用SURF进行手势到文本的翻译

IF 3.8 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Applied System Innovation Pub Date : 2023-03-02 DOI:10.3390/asi6020035
Kaustubh Mani Tripathi, P. Kamat, S. Patil, Ruchi Jayaswal, Swati Ahirrao, K. Kotecha
{"title":"用SURF进行手势到文本的翻译","authors":"Kaustubh Mani Tripathi, P. Kamat, S. Patil, Ruchi Jayaswal, Swati Ahirrao, K. Kotecha","doi":"10.3390/asi6020035","DOIUrl":null,"url":null,"abstract":"This research paper focuses on developing an effective gesture-to-text translation system using state-of-the-art computer vision techniques. The existing research on sign language translation has yet to utilize skin masking, edge detection, and feature extraction techniques to their full potential. Therefore, this study employs the speeded-up robust features (SURF) model for feature extraction, which is resistant to variations such as rotation, perspective scaling, and occlusion. The proposed system utilizes a bag of visual words (BoVW) model for gesture-to-text conversion. The study uses a dataset of 42,000 photographs consisting of alphabets (A–Z) and numbers (1–9), divided into 35 classes with 1200 shots per class. The pre-processing phase includes skin masking, where the RGB color space is converted to the HSV color space, and Canny edge detection is used for sharp edge detection. The SURF elements are grouped and converted to a visual language using the K-means mini-batch clustering technique. The proposed system’s performance is evaluated using several machine learning algorithms such as naïve Bayes, logistic regression, K nearest neighbors, support vector machine, and convolutional neural network. All the algorithms benefited from SURF, and the system’s accuracy is promising, ranging from 79% to 92%. This research study not only presents the development of an effective gesture-to-text translation system but also highlights the importance of using skin masking, edge detection, and feature extraction techniques to their full potential in sign language translation. The proposed system aims to bridge the communication gap between individuals who cannot speak and those who cannot understand Indian Sign Language (ISL).","PeriodicalId":36273,"journal":{"name":"Applied System Innovation","volume":" ","pages":""},"PeriodicalIF":3.8000,"publicationDate":"2023-03-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Gesture-to-Text Translation Using SURF for Indian Sign Language\",\"authors\":\"Kaustubh Mani Tripathi, P. Kamat, S. Patil, Ruchi Jayaswal, Swati Ahirrao, K. Kotecha\",\"doi\":\"10.3390/asi6020035\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This research paper focuses on developing an effective gesture-to-text translation system using state-of-the-art computer vision techniques. The existing research on sign language translation has yet to utilize skin masking, edge detection, and feature extraction techniques to their full potential. Therefore, this study employs the speeded-up robust features (SURF) model for feature extraction, which is resistant to variations such as rotation, perspective scaling, and occlusion. The proposed system utilizes a bag of visual words (BoVW) model for gesture-to-text conversion. The study uses a dataset of 42,000 photographs consisting of alphabets (A–Z) and numbers (1–9), divided into 35 classes with 1200 shots per class. The pre-processing phase includes skin masking, where the RGB color space is converted to the HSV color space, and Canny edge detection is used for sharp edge detection. The SURF elements are grouped and converted to a visual language using the K-means mini-batch clustering technique. The proposed system’s performance is evaluated using several machine learning algorithms such as naïve Bayes, logistic regression, K nearest neighbors, support vector machine, and convolutional neural network. All the algorithms benefited from SURF, and the system’s accuracy is promising, ranging from 79% to 92%. This research study not only presents the development of an effective gesture-to-text translation system but also highlights the importance of using skin masking, edge detection, and feature extraction techniques to their full potential in sign language translation. The proposed system aims to bridge the communication gap between individuals who cannot speak and those who cannot understand Indian Sign Language (ISL).\",\"PeriodicalId\":36273,\"journal\":{\"name\":\"Applied System Innovation\",\"volume\":\" \",\"pages\":\"\"},\"PeriodicalIF\":3.8000,\"publicationDate\":\"2023-03-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Applied System Innovation\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.3390/asi6020035\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied System Innovation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3390/asi6020035","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0

摘要

本文的研究重点是利用最先进的计算机视觉技术开发一个有效的手势到文本的翻译系统。现有的手语翻译研究尚未充分利用皮肤掩蔽、边缘检测和特征提取技术。因此,本研究采用了加速鲁棒特征(SURF)模型进行特征提取,该模型能够抵抗旋转、透视缩放和遮挡等变化。所提出的系统利用视觉单词袋(BoVW)模型进行手势到文本的转换。该研究使用了一个由42000张照片组成的数据集,这些照片由字母(a-Z)和数字(1-9)组成,分为35个类别,每个类别1200张照片。预处理阶段包括皮肤掩蔽,其中RGB颜色空间被转换为HSV颜色空间,并且Canny边缘检测用于尖锐边缘检测。SURF元素被分组,并使用K-means小批量聚类技术转换为视觉语言。使用几种机器学习算法,如朴素贝叶斯、逻辑回归、K近邻、支持向量机和卷积神经网络,对所提出的系统的性能进行了评估。所有算法都受益于SURF,系统的准确率很有希望,从79%到92%不等。这项研究不仅介绍了一种有效的手势到文本翻译系统的开发,还强调了使用皮肤掩蔽、边缘检测和特征提取技术在手语翻译中充分发挥潜力的重要性。拟议的系统旨在弥合不会说话的人和不懂印度手语的人之间的沟通差距。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Gesture-to-Text Translation Using SURF for Indian Sign Language
This research paper focuses on developing an effective gesture-to-text translation system using state-of-the-art computer vision techniques. The existing research on sign language translation has yet to utilize skin masking, edge detection, and feature extraction techniques to their full potential. Therefore, this study employs the speeded-up robust features (SURF) model for feature extraction, which is resistant to variations such as rotation, perspective scaling, and occlusion. The proposed system utilizes a bag of visual words (BoVW) model for gesture-to-text conversion. The study uses a dataset of 42,000 photographs consisting of alphabets (A–Z) and numbers (1–9), divided into 35 classes with 1200 shots per class. The pre-processing phase includes skin masking, where the RGB color space is converted to the HSV color space, and Canny edge detection is used for sharp edge detection. The SURF elements are grouped and converted to a visual language using the K-means mini-batch clustering technique. The proposed system’s performance is evaluated using several machine learning algorithms such as naïve Bayes, logistic regression, K nearest neighbors, support vector machine, and convolutional neural network. All the algorithms benefited from SURF, and the system’s accuracy is promising, ranging from 79% to 92%. This research study not only presents the development of an effective gesture-to-text translation system but also highlights the importance of using skin masking, edge detection, and feature extraction techniques to their full potential in sign language translation. The proposed system aims to bridge the communication gap between individuals who cannot speak and those who cannot understand Indian Sign Language (ISL).
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Applied System Innovation
Applied System Innovation Mathematics-Applied Mathematics
CiteScore
7.90
自引率
5.30%
发文量
102
审稿时长
11 weeks
期刊最新文献
Research on Chinese Nested Entity Recognition Based on IDCNNLR and GlobalPointer AI-Powered Academic Guidance and Counseling System Based on Student Profile and Interests Using Smart Traffic Lights to Reduce CO2 Emissions and Improve Traffic Flow at Intersections: Simulation of an Intersection in a Small Portuguese City Predictive Modeling of Light–Matter Interaction in One Dimension: A Dynamic Deep Learning Approach Project Management Efficiency Measurement with Data Envelopment Analysis: A Case in a Petrochemical Company
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1