首页 > 最新文献

International Conference on Artificial Intelligence, Virtual Reality, and Visualization最新文献

英文 中文
Real-time visual monitoring system and method of coal spontaneous combustion temperature field in goaf area 采空区煤自燃温度场实时可视化监测系统与方法
Zhen Xing
A visual monitoring equipment for coal spontaneous combustion temperature field in goaf is designed,The automatic operation system and self-cleaning device are designed according to the coal mine goaf site,An optimal solution algorithm for temperature anomaly area in goaf is developed and embedded software is formed,The collected data are analyzed through the model to obtain the optimal location of high temperature ignition source in the goaf, and the data are uploaded to the monitoring center through the transmission network.The system formed by the combination of the device and determination method can make visual monitoring and intelligent judgment of the danger area in goaf.
设计了采空区煤炭自燃温度场可视化监测设备,根据煤矿采空区现场设计了自动操作系统和自清洗装置,开发了采空区温度异常区的最优解算法,并形成了嵌入式软件。通过模型对采集到的数据进行分析,得到采空区高温点火源的最佳位置,并通过传输网络将数据上传到监测中心。该装置与确定方法相结合形成的系统,可以对采空区危险区域进行可视化监测和智能判断。
{"title":"Real-time visual monitoring system and method of coal spontaneous combustion temperature field in goaf area","authors":"Zhen Xing","doi":"10.1117/12.2667194","DOIUrl":"https://doi.org/10.1117/12.2667194","url":null,"abstract":"A visual monitoring equipment for coal spontaneous combustion temperature field in goaf is designed,The automatic operation system and self-cleaning device are designed according to the coal mine goaf site,An optimal solution algorithm for temperature anomaly area in goaf is developed and embedded software is formed,The collected data are analyzed through the model to obtain the optimal location of high temperature ignition source in the goaf, and the data are uploaded to the monitoring center through the transmission network.The system formed by the combination of the device and determination method can make visual monitoring and intelligent judgment of the danger area in goaf.","PeriodicalId":137914,"journal":{"name":"International Conference on Artificial Intelligence, Virtual Reality, and Visualization","volume":"80 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128383851","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Harnessing transfer learning for Alzheimer's disease prediction 利用迁移学习预测阿尔茨海默病
Yukun Liu, Chengxuan Zheng, Baha Ihnaini
Nowadays, Alzheimer's Disease (AD) has become a massive problem for middle-aged and older adults. Although due to its long incubation period and early mild symptoms, patients have a more extended period and more possibilities to check out, it is still hard for patients and doctors to diagnose in early routine examinations. This article provides a new method to help the doctor to diagnose Alzheimer's Disease in the early phase. We use transfer learning in deep learning to help diagnose Alzheimer's Disease early in developing Computed Tomography (CT) brain images. Using three pre-trained models, ShuffleNet, DenseNet, and NASNet-mobile as the transfer learning training model and convolution neural networks. We made some improvements to make it more relevant to the actual situation. DenseNet has best performance (87.36%) among the three models. We set the output into four classes: the four stages of Alzheimer's are widely recognized (Mild Demented, Moderate Demented, Very Mild Demented).
如今,阿尔茨海默病(AD)已成为中老年人的一大问题。虽然由于潜伏期长,早期症状轻,患者有更长的时间和更多的检查可能性,但在早期常规检查中,患者和医生仍然难以诊断。本文为帮助医生早期诊断阿尔茨海默病提供了一种新的方法。我们在深度学习中使用迁移学习来帮助在开发计算机断层扫描(CT)脑图像的早期诊断阿尔茨海默病。使用三个预训练模型,ShuffleNet, DenseNet和NASNet-mobile作为迁移学习训练模型和卷积神经网络。我们做了一些改进,使之更贴近实际情况。DenseNet在3种模型中表现最好(87.36%)。我们将输出分为四类:阿尔茨海默氏症的四个阶段被广泛认可(轻度痴呆,中度痴呆,非常轻度痴呆)。
{"title":"Harnessing transfer learning for Alzheimer's disease prediction","authors":"Yukun Liu, Chengxuan Zheng, Baha Ihnaini","doi":"10.1117/12.2667247","DOIUrl":"https://doi.org/10.1117/12.2667247","url":null,"abstract":"Nowadays, Alzheimer's Disease (AD) has become a massive problem for middle-aged and older adults. Although due to its long incubation period and early mild symptoms, patients have a more extended period and more possibilities to check out, it is still hard for patients and doctors to diagnose in early routine examinations. This article provides a new method to help the doctor to diagnose Alzheimer's Disease in the early phase. We use transfer learning in deep learning to help diagnose Alzheimer's Disease early in developing Computed Tomography (CT) brain images. Using three pre-trained models, ShuffleNet, DenseNet, and NASNet-mobile as the transfer learning training model and convolution neural networks. We made some improvements to make it more relevant to the actual situation. DenseNet has best performance (87.36%) among the three models. We set the output into four classes: the four stages of Alzheimer's are widely recognized (Mild Demented, Moderate Demented, Very Mild Demented).","PeriodicalId":137914,"journal":{"name":"International Conference on Artificial Intelligence, Virtual Reality, and Visualization","volume":"698 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132780883","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Visualization of planning spaces in virtual reality 虚拟现实中规划空间的可视化
Haiwei Zuo, Wanghui Chu, Bo Yang
The use of virtual reality (VR) in urban area planning is examined in this dissertation. There are numerous allusions to Openstreetmap's urban building data. This VR technology allows users to enter a virtual environment using an input device (HTC Vive). Without actually carrying out the plan, the effectiveness of urban planning can be tested beforehand. VR will become more crucial in urban area planning, boosting urban construction, and assisting critical individuals in making decisions. There were three primary tasks for this project: First task: Create a 3D urban model using the OpenStreetMap map file. Unreal Engine 4 has imported a file with an OSM map (UE4). A data stream has been created to convert OSM files to UE4 uasset files, according to StreepMap and the RuntimeMeshComponent plugin. Second task: Create an overview mode. Users can highlight various items with various features on the overview map in this mode, including different building levels, the top speed limit, hotel star ratings, amenity categories, and retail categories. Third task: Create a VR mode. A virtual environment has been developed to enhance immersion and user experience. The HTC VIVE controller allows users to explore city maps independently. A new interactive interface has been created for this project. With the aid of this method, urban planning initiatives can be pretested before being put into action. Contributions to science: This project's main contribution is developing an application to aid pertinent staff members in managing and planning urban data. Based on this research, urban spatial planning could cut costs by using virtual reality 3D modeling, data integration, and VR interaction.
本文对虚拟现实技术在城市规划中的应用进行了研究。Openstreetmap的城市建筑数据有很多典故。这种虚拟现实技术允许用户使用输入设备(HTC Vive)进入虚拟环境。在不实际执行规划的情况下,可以事先测试城市规划的有效性。虚拟现实将在城市规划、推动城市建设和协助关键个人决策方面发挥更重要的作用。这个项目有三个主要任务:第一个任务:使用OpenStreetMap地图文件创建一个3D城市模型。虚幻引擎4已经导入了一个带有OSM映射(UE4)的文件。根据StreepMap和RuntimeMeshComponent插件,已经创建了一个数据流来将OSM文件转换为UE4 uasset文件。第二个任务:创建概览模式。在这种模式下,用户可以在概览地图上突出显示具有各种功能的各种项目,包括不同的建筑级别、最高速度限制、酒店星级评级、便利设施类别和零售类别。第三个任务:创建VR模式。一个虚拟环境已经被开发出来,以增强沉浸感和用户体验。HTC VIVE控制器允许用户独立探索城市地图。为此项目创建了一个新的交互界面。在这种方法的帮助下,城市规划倡议可以在付诸行动之前进行预测试。对科学的贡献:该项目的主要贡献是开发一个应用程序,以帮助相关工作人员管理和规划城市数据。基于本研究,城市空间规划可以通过虚拟现实三维建模、数据集成和VR交互来降低成本。
{"title":"Visualization of planning spaces in virtual reality","authors":"Haiwei Zuo, Wanghui Chu, Bo Yang","doi":"10.1117/12.2667202","DOIUrl":"https://doi.org/10.1117/12.2667202","url":null,"abstract":"The use of virtual reality (VR) in urban area planning is examined in this dissertation. There are numerous allusions to Openstreetmap's urban building data. This VR technology allows users to enter a virtual environment using an input device (HTC Vive). Without actually carrying out the plan, the effectiveness of urban planning can be tested beforehand. VR will become more crucial in urban area planning, boosting urban construction, and assisting critical individuals in making decisions. There were three primary tasks for this project: First task: Create a 3D urban model using the OpenStreetMap map file. Unreal Engine 4 has imported a file with an OSM map (UE4). A data stream has been created to convert OSM files to UE4 uasset files, according to StreepMap and the RuntimeMeshComponent plugin. Second task: Create an overview mode. Users can highlight various items with various features on the overview map in this mode, including different building levels, the top speed limit, hotel star ratings, amenity categories, and retail categories. Third task: Create a VR mode. A virtual environment has been developed to enhance immersion and user experience. The HTC VIVE controller allows users to explore city maps independently. A new interactive interface has been created for this project. With the aid of this method, urban planning initiatives can be pretested before being put into action. Contributions to science: This project's main contribution is developing an application to aid pertinent staff members in managing and planning urban data. Based on this research, urban spatial planning could cut costs by using virtual reality 3D modeling, data integration, and VR interaction.","PeriodicalId":137914,"journal":{"name":"International Conference on Artificial Intelligence, Virtual Reality, and Visualization","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122002911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research and application of template matching algorithm based on edge contour 基于边缘轮廓的模板匹配算法研究与应用
Zhuoxin Liu, Ji Zhang, Kaibin Chu
In this paper, the ideas of shape template matching, multi-sample template matching, and ROI area processing are combined into an edge contour template matching algorithm module. Based on the OpenCV open source vision library, the area to be removed in the ROI area is colored polygons, and the noise points of this shape feature need to be removed, and only the feature points with obvious edge contour features are extracted; as a result, the matching time is transferred to the mapping time, thereby reducing the total matching time.
本文将形状模板匹配、多样本模板匹配、ROI区域处理等思想结合到边缘轮廓模板匹配算法模块中。基于OpenCV开源视觉库,ROI区域中待去除的区域为彩色多边形,需要去除该形状特征的噪声点,只提取边缘轮廓特征明显的特征点;因此,将匹配时间转移到映射时间,从而减少了总匹配时间。
{"title":"Research and application of template matching algorithm based on edge contour","authors":"Zhuoxin Liu, Ji Zhang, Kaibin Chu","doi":"10.1117/12.2667677","DOIUrl":"https://doi.org/10.1117/12.2667677","url":null,"abstract":"In this paper, the ideas of shape template matching, multi-sample template matching, and ROI area processing are combined into an edge contour template matching algorithm module. Based on the OpenCV open source vision library, the area to be removed in the ROI area is colored polygons, and the noise points of this shape feature need to be removed, and only the feature points with obvious edge contour features are extracted; as a result, the matching time is transferred to the mapping time, thereby reducing the total matching time.","PeriodicalId":137914,"journal":{"name":"International Conference on Artificial Intelligence, Virtual Reality, and Visualization","volume":"341 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115473964","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semantic enhancement methods for image captioning 图像字幕的语义增强方法
Luming Cui, Lin Li
Image captioning, a cross-modal study, aims to generating a description for a given image, which plays an important role in many fields like image retrieval and computer-assisted instruction. Currently, the challenge in image captioning is the limited quality of generated descriptions including insufficient utilization of image feature information and the limited language learning ability of the decoder. In this paper, we address the above problems by constructing a semantic enhancement module and a multi-round decoding mechanism to enhance the decoding ability of the model, which uses the Transformer model as the primary structure. To validate the efficacy of the model, we conducted intensive experiments on the MSCOCO2014 benchmark and evaluated its performance using five evaluation metrics. The experimental results show that the proposed method in this paper has improved to varying degrees on all five-evaluation metrics.
图像字幕是一种跨模态的研究,其目的是为给定的图像生成描述,在图像检索和计算机辅助教学等许多领域发挥着重要作用。目前,图像字幕所面临的挑战是生成的描述质量有限,包括图像特征信息的利用不足以及解码器的语言学习能力有限。本文以Transformer模型为主要结构,通过构建语义增强模块和多轮解码机制来提高模型的解码能力,从而解决了上述问题。为了验证模型的有效性,我们在MSCOCO2014基准上进行了大量实验,并使用五个评价指标对其性能进行了评估。实验结果表明,本文提出的方法在五个评价指标上都有不同程度的改进。
{"title":"Semantic enhancement methods for image captioning","authors":"Luming Cui, Lin Li","doi":"10.1117/12.2667270","DOIUrl":"https://doi.org/10.1117/12.2667270","url":null,"abstract":"Image captioning, a cross-modal study, aims to generating a description for a given image, which plays an important role in many fields like image retrieval and computer-assisted instruction. Currently, the challenge in image captioning is the limited quality of generated descriptions including insufficient utilization of image feature information and the limited language learning ability of the decoder. In this paper, we address the above problems by constructing a semantic enhancement module and a multi-round decoding mechanism to enhance the decoding ability of the model, which uses the Transformer model as the primary structure. To validate the efficacy of the model, we conducted intensive experiments on the MSCOCO2014 benchmark and evaluated its performance using five evaluation metrics. The experimental results show that the proposed method in this paper has improved to varying degrees on all five-evaluation metrics.","PeriodicalId":137914,"journal":{"name":"International Conference on Artificial Intelligence, Virtual Reality, and Visualization","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115351070","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An enhanced grammatical approach for graph drawing 图形绘制的增强语法方法
Yufeng Liu, Yangchen Zhou, Fan Yang, Hongliang Sun
With the development of computer-aided design, visual languages have been widely used as formal methods to represent various types of graphical models. Accordingly, many grammar systems have been proposed for the specification of visual languages. However, apart from shape grammar, most grammars focus on the abstract structures of the languages and ignore the semantic modeling of graph drawing. Furthermore, shape grammar supports generation rather than parsing, with its limited application scope. To address these problems, this paper proposes an enhanced grammar system based on Coordinate Graph Grammar (CGG). Different from traditional grammars, the enhanced system defines a new type of grammatical rule named shape rules to transform graphs into shapes by shape applications. In each shape application, the assertion set describes the range of validity, and shapes can be generated by translation, zoom, and rotation to a set of rule-based coordinates. With the combinations of shape applications and L-applications, the node-edge graph and drawn outline could both be specified, building a bridge between abstract structures and physical layouts of visual languages. An example is given to illustrate the application of the enhanced system in industrial design, where a Bauhaus-style baby cradle is generated by the combination of shape applications and L-applications.
随着计算机辅助设计技术的发展,视觉语言作为一种形式化的方法被广泛应用于各种类型的图形模型的表示。因此,人们提出了许多语法系统来规范视觉语言。然而,除了形状语法外,大多数语法都侧重于语言的抽象结构,而忽略了图形绘制的语义建模。此外,形状语法支持生成而不是解析,其应用范围有限。针对这些问题,本文提出了一种基于坐标图语法(CGG)的增强型语法体系。与传统语法不同的是,增强后的系统定义了一种新的语法规则——形状规则,通过形状应用将图形转换为形状。在每个形状应用程序中,断言集描述有效范围,形状可以通过转换、缩放和旋转到一组基于规则的坐标来生成。通过形状应用和l应用的结合,可以同时指定节点边缘图和绘制的轮廓,在抽象结构和视觉语言的物理布局之间架起一座桥梁。最后给出了该增强系统在工业设计中的应用实例,将形状应用与l型应用相结合,生成了包豪斯风格的婴儿摇篮。
{"title":"An enhanced grammatical approach for graph drawing","authors":"Yufeng Liu, Yangchen Zhou, Fan Yang, Hongliang Sun","doi":"10.1117/12.2667201","DOIUrl":"https://doi.org/10.1117/12.2667201","url":null,"abstract":"With the development of computer-aided design, visual languages have been widely used as formal methods to represent various types of graphical models. Accordingly, many grammar systems have been proposed for the specification of visual languages. However, apart from shape grammar, most grammars focus on the abstract structures of the languages and ignore the semantic modeling of graph drawing. Furthermore, shape grammar supports generation rather than parsing, with its limited application scope. To address these problems, this paper proposes an enhanced grammar system based on Coordinate Graph Grammar (CGG). Different from traditional grammars, the enhanced system defines a new type of grammatical rule named shape rules to transform graphs into shapes by shape applications. In each shape application, the assertion set describes the range of validity, and shapes can be generated by translation, zoom, and rotation to a set of rule-based coordinates. With the combinations of shape applications and L-applications, the node-edge graph and drawn outline could both be specified, building a bridge between abstract structures and physical layouts of visual languages. An example is given to illustrate the application of the enhanced system in industrial design, where a Bauhaus-style baby cradle is generated by the combination of shape applications and L-applications.","PeriodicalId":137914,"journal":{"name":"International Conference on Artificial Intelligence, Virtual Reality, and Visualization","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122903008","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research on identity authentication and labeling technology based on MR neural network 基于MR神经网络的身份认证与标识技术研究
Hao Yang, Chuan-qian Tang
Aiming at the problems of poor convenience, poor scalability, and low authentication rate in traditional authentication technology using physical contact authentication methods such as magnetic cards and passwords, this paper explores the accuracy and convenience of the practical application of MR neural network in personal identity authentication. In the MR wearable device, the neural network person identity authentication method is studied flexibly and quickly to detect and identify the person. The 3D information of the face is collected and preprocessed by the depth camera, and the MR identity authentication data set is established. The neural network Resnet model is used for face detection and face feature vector extraction, and the Euclidean method is used to compare the feature vectors and label the characters. The neural network authentication algorithm is mapped to the MR wearable device, and the deep face information in the scene is identified, matched, and labeled by using the unique spatial mapping of MR technology and the camera of the MR wearable device. It solves the problems of low flexibility, poor reliability of face information, and weak recognition stability in traditional identity authentication methods, enabling MR technology to provide a more intelligent identification and labeling method for person identity authentication.
针对传统的磁卡、密码等物理接触认证方式的认证技术存在便捷性差、可扩展性差、认证率低等问题,探讨MR神经网络在个人身份认证中实际应用的准确性和便捷性。在MR可穿戴设备中,研究了灵活快速的神经网络人身份认证方法来检测和识别人。通过深度相机采集人脸三维信息并进行预处理,建立MR身份认证数据集。采用神经网络Resnet模型进行人脸检测和人脸特征向量提取,并采用欧几里得方法进行特征向量比较和字符标注。将神经网络认证算法映射到MR可穿戴设备上,利用MR技术独特的空间映射和MR可穿戴设备的摄像头,对场景中的深层人脸信息进行识别、匹配和标记。它解决了传统身份认证方法中人脸信息灵活性低、可靠性差、识别稳定性弱的问题,使MR技术能够为人的身份认证提供更加智能的识别和标注方法。
{"title":"Research on identity authentication and labeling technology based on MR neural network","authors":"Hao Yang, Chuan-qian Tang","doi":"10.1117/12.2667283","DOIUrl":"https://doi.org/10.1117/12.2667283","url":null,"abstract":"Aiming at the problems of poor convenience, poor scalability, and low authentication rate in traditional authentication technology using physical contact authentication methods such as magnetic cards and passwords, this paper explores the accuracy and convenience of the practical application of MR neural network in personal identity authentication. In the MR wearable device, the neural network person identity authentication method is studied flexibly and quickly to detect and identify the person. The 3D information of the face is collected and preprocessed by the depth camera, and the MR identity authentication data set is established. The neural network Resnet model is used for face detection and face feature vector extraction, and the Euclidean method is used to compare the feature vectors and label the characters. The neural network authentication algorithm is mapped to the MR wearable device, and the deep face information in the scene is identified, matched, and labeled by using the unique spatial mapping of MR technology and the camera of the MR wearable device. It solves the problems of low flexibility, poor reliability of face information, and weak recognition stability in traditional identity authentication methods, enabling MR technology to provide a more intelligent identification and labeling method for person identity authentication.","PeriodicalId":137914,"journal":{"name":"International Conference on Artificial Intelligence, Virtual Reality, and Visualization","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131596935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Measurement error detection method of electric energy meter based on machine vision 基于机器视觉的电能表测量误差检测方法
Zhen Gu, Da-rong Chen, J. Wang, Chen Dai, Gewei Zhuang
Due to the slow response and poor accuracy of traditional measurement error detection of the electric energy meter, the measurement error detection method of electric energy meter based on machine vision is studied. The minimum error method is used to segment the image threshold to form a binary image. The morphological refinement method is used to extract the image edge contour, combined with machine vision to refine the edge pixels, to achieve the measurement error detection of the instrument. The experimental results show that using the error detection method of machine vision, the detection results are consistent with the error detection results set by the system and the trend is the same. The accuracy also meets the requirements of relevant regulations, which improves the accuracy of electric energy meter measurement.
针对传统电能表测量误差检测响应慢、精度差的问题,研究了基于机器视觉的电能表测量误差检测方法。采用最小误差法对图像阈值进行分割,形成二值图像。采用形态学细化方法提取图像边缘轮廓,结合机器视觉对边缘像素进行细化,实现仪器的测量误差检测。实验结果表明,采用机器视觉的误差检测方法,检测结果与系统设置的误差检测结果一致,趋势一致。精度也符合相关规定的要求,提高了电能表计量的精度。
{"title":"Measurement error detection method of electric energy meter based on machine vision","authors":"Zhen Gu, Da-rong Chen, J. Wang, Chen Dai, Gewei Zhuang","doi":"10.1117/12.2667641","DOIUrl":"https://doi.org/10.1117/12.2667641","url":null,"abstract":"Due to the slow response and poor accuracy of traditional measurement error detection of the electric energy meter, the measurement error detection method of electric energy meter based on machine vision is studied. The minimum error method is used to segment the image threshold to form a binary image. The morphological refinement method is used to extract the image edge contour, combined with machine vision to refine the edge pixels, to achieve the measurement error detection of the instrument. The experimental results show that using the error detection method of machine vision, the detection results are consistent with the error detection results set by the system and the trend is the same. The accuracy also meets the requirements of relevant regulations, which improves the accuracy of electric energy meter measurement.","PeriodicalId":137914,"journal":{"name":"International Conference on Artificial Intelligence, Virtual Reality, and Visualization","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131735876","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Typical wire clamps segmentation of transmission lines based on infrared image 基于红外图像的传输线典型线钳分割
Ding-Kuo Huang, Jie Yang, Yunfeng Yan, Xiangwei Sun, Xiaoming Huang
In view of the current image segmentation field, there are few studies on the segmentation of typical wire clamp components of transmission lines. Traditional image processing methods have low segmentation accuracy and require artificial design of feature extraction methods, which are usually only suitable for equipment of a certain structure with insufficient generalization. In this paper, an infrared image segmentation method based on Mask R-CNN (Mask region-based convolutional neural network) for typical guide-ground lines is proposed. Its structure takes Mask R-CNN model combined with FPN (Feature pyramid structure) as the basic framework, and uses RPN (Regional proposal network) to generate candidate regions. Features are extracted from each candidate region through RoI Align layer, and then connected to FC (Fully connected layer) to achieve target classification and bbox (bounding box) regression. A mask branch is also added to predict the segmentation mask. The design can integrate multi-scale and multi-level semantic information to improve the recognition rate when extracting image features. In addition, the network structure is optimized by single channel for infrared images to reduce the size of the model and make it more lightweight. Ablation experiments were performed on two GTX 2080Ti graphics cards to verify the effectiveness of the proposed structure, and the mAP (mean average accuracy) of 0.421 was achieved with an IoU (Intersection over Union) threshold of 0.5.
针对目前的图像分割领域,对传输线典型线钳部件的分割研究较少。传统的图像处理方法分割精度低,需要人工设计特征提取方法,通常只适用于某种结构的设备,泛化程度不够。本文提出了一种基于Mask - cnn (Mask region-based convolutional neural network)的典型导地线红外图像分割方法。其结构以Mask R-CNN模型结合FPN (Feature pyramid structure)作为基本框架,使用RPN (Regional proposal network)生成候选区域。通过RoI Align层提取每个候选区域的特征,然后与FC (Fully connected layer)连接,实现目标分类和bbox (bounding box)回归。还添加了一个掩码分支来预测分割掩码。该设计能够融合多尺度、多层次的语义信息,提高图像特征提取的识别率。此外,针对红外图像采用单通道优化网络结构,减小模型尺寸,使其更轻量化。在两张GTX 2080Ti显卡上进行了消融实验,验证了所提出结构的有效性,在IoU (Intersection over Union)阈值为0.5的情况下,mAP(平均精度)达到0.421。
{"title":"Typical wire clamps segmentation of transmission lines based on infrared image","authors":"Ding-Kuo Huang, Jie Yang, Yunfeng Yan, Xiangwei Sun, Xiaoming Huang","doi":"10.1117/12.2667507","DOIUrl":"https://doi.org/10.1117/12.2667507","url":null,"abstract":"In view of the current image segmentation field, there are few studies on the segmentation of typical wire clamp components of transmission lines. Traditional image processing methods have low segmentation accuracy and require artificial design of feature extraction methods, which are usually only suitable for equipment of a certain structure with insufficient generalization. In this paper, an infrared image segmentation method based on Mask R-CNN (Mask region-based convolutional neural network) for typical guide-ground lines is proposed. Its structure takes Mask R-CNN model combined with FPN (Feature pyramid structure) as the basic framework, and uses RPN (Regional proposal network) to generate candidate regions. Features are extracted from each candidate region through RoI Align layer, and then connected to FC (Fully connected layer) to achieve target classification and bbox (bounding box) regression. A mask branch is also added to predict the segmentation mask. The design can integrate multi-scale and multi-level semantic information to improve the recognition rate when extracting image features. In addition, the network structure is optimized by single channel for infrared images to reduce the size of the model and make it more lightweight. Ablation experiments were performed on two GTX 2080Ti graphics cards to verify the effectiveness of the proposed structure, and the mAP (mean average accuracy) of 0.421 was achieved with an IoU (Intersection over Union) threshold of 0.5.","PeriodicalId":137914,"journal":{"name":"International Conference on Artificial Intelligence, Virtual Reality, and Visualization","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134115249","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Heterogeneous models ensemble for Chinese grammatical error correction 汉语语法错误纠错的异构模型集成
Yeling Liang, Lin Li
Grammatical error correction (GEC) aims to automatically identify and correct grammatical errors in a sentence. Neural machine translation (NMT) models are the mainstream approaches for the GEC task. However, the models require a large amount of data to be adequately trained, the variety of grammatical errors and the dependencies between errors in a sentence make it difficult for a single NMT model to correct multiple errors at once. In the work, we propose an ensemble approach for heterogeneous models, which integrates rule-based, NMT, and pre-trained language model-based GEC models through the recurrent generation approach, the approach can exploit the strengths of each model and cover a wider range of errors in a sentence. We also mitigate the scarcity of task-specific data for the GEC task through the data augmentation approach. We conduct extensive experiments on the NLPCC2018 shared task dataset to demonstrate the effectiveness of our proposed methods, and reaches the F0.5 value of 37.26, outperforming the best model in the shared task.
语法错误纠正(GEC)旨在自动识别和纠正句子中的语法错误。神经机器翻译(NMT)模型是解决GEC任务的主流方法。然而,这些模型需要大量的数据进行充分的训练,语法错误的多样性以及句子中错误之间的依赖性使得单个NMT模型很难同时纠正多个错误。在这项工作中,我们提出了一种异构模型的集成方法,该方法通过循环生成方法集成了基于规则的、NMT的和基于预训练语言模型的GEC模型,该方法可以利用每个模型的优势,并覆盖句子中更大范围的错误。我们还通过数据增强方法缓解了GEC任务特定于任务的数据的稀缺性。我们在NLPCC2018共享任务数据集上进行了大量的实验,证明了我们提出的方法的有效性,F0.5值达到37.26,优于共享任务中的最佳模型。
{"title":"Heterogeneous models ensemble for Chinese grammatical error correction","authors":"Yeling Liang, Lin Li","doi":"10.1117/12.2667512","DOIUrl":"https://doi.org/10.1117/12.2667512","url":null,"abstract":"Grammatical error correction (GEC) aims to automatically identify and correct grammatical errors in a sentence. Neural machine translation (NMT) models are the mainstream approaches for the GEC task. However, the models require a large amount of data to be adequately trained, the variety of grammatical errors and the dependencies between errors in a sentence make it difficult for a single NMT model to correct multiple errors at once. In the work, we propose an ensemble approach for heterogeneous models, which integrates rule-based, NMT, and pre-trained language model-based GEC models through the recurrent generation approach, the approach can exploit the strengths of each model and cover a wider range of errors in a sentence. We also mitigate the scarcity of task-specific data for the GEC task through the data augmentation approach. We conduct extensive experiments on the NLPCC2018 shared task dataset to demonstrate the effectiveness of our proposed methods, and reaches the F0.5 value of 37.26, outperforming the best model in the shared task.","PeriodicalId":137914,"journal":{"name":"International Conference on Artificial Intelligence, Virtual Reality, and Visualization","volume":"205 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114649217","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
International Conference on Artificial Intelligence, Virtual Reality, and Visualization
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1