帮助视障人士的印度货币识别模型

Madhav Pasumarthy, Rutvi Padhy, Raghuveer Yadav, Ganesh Subramaniam, Madhav Rao
{"title":"帮助视障人士的印度货币识别模型","authors":"Madhav Pasumarthy, Rutvi Padhy, Raghuveer Yadav, Ganesh Subramaniam, Madhav Rao","doi":"10.1109/RASSE54974.2022.9989624","DOIUrl":null,"url":null,"abstract":"Visually impaired persons find it extremely difficult to perform cash transactions in outdoor environments. For assisting the visually challenged individuals, a YOLOv5 based deep neural network was designed to detect image based currency denominations. Thereby aid in completing the authentic transaction. The robust model was trained for images with currency notes in different backgrounds, multiple sides of the currency notes presented, notes around cluttered objects, notes near reflective surfaces, and blurred images of the currency notes. An annotated and augmented dataset of around 10,000 original images was created for developing the model. A pre-processing step to rescale all the images to 224 × 224 was applied to standardize the input to the neural network, and generalize the model for different platforms including single board computer and smartphones. The trained model showcased an average denomination recognition accuracy of 92.71% for an altogether different dataset. The trained model was deployed on Raspberry-Pi and Smartphone independently, and the outcome to detect the currency denomination from the image was successfully demonstrated. The model showcased adequate performance on different platforms, leading to the exploration of several other assistive applications based on the currency recognition model to improve the standard of living for visually challenged individuals.","PeriodicalId":382440,"journal":{"name":"2022 IEEE International Conference on Recent Advances in Systems Science and Engineering (RASSE)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"An Indian Currency Recognition Model for Assisting Visually Impaired Individuals\",\"authors\":\"Madhav Pasumarthy, Rutvi Padhy, Raghuveer Yadav, Ganesh Subramaniam, Madhav Rao\",\"doi\":\"10.1109/RASSE54974.2022.9989624\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Visually impaired persons find it extremely difficult to perform cash transactions in outdoor environments. For assisting the visually challenged individuals, a YOLOv5 based deep neural network was designed to detect image based currency denominations. Thereby aid in completing the authentic transaction. The robust model was trained for images with currency notes in different backgrounds, multiple sides of the currency notes presented, notes around cluttered objects, notes near reflective surfaces, and blurred images of the currency notes. An annotated and augmented dataset of around 10,000 original images was created for developing the model. A pre-processing step to rescale all the images to 224 × 224 was applied to standardize the input to the neural network, and generalize the model for different platforms including single board computer and smartphones. The trained model showcased an average denomination recognition accuracy of 92.71% for an altogether different dataset. The trained model was deployed on Raspberry-Pi and Smartphone independently, and the outcome to detect the currency denomination from the image was successfully demonstrated. The model showcased adequate performance on different platforms, leading to the exploration of several other assistive applications based on the currency recognition model to improve the standard of living for visually challenged individuals.\",\"PeriodicalId\":382440,\"journal\":{\"name\":\"2022 IEEE International Conference on Recent Advances in Systems Science and Engineering (RASSE)\",\"volume\":\"6 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-11-07\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE International Conference on Recent Advances in Systems Science and Engineering (RASSE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/RASSE54974.2022.9989624\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE International Conference on Recent Advances in Systems Science and Engineering (RASSE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/RASSE54974.2022.9989624","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

视障人士发现在户外环境下进行现金交易极为困难。为了帮助视障人士,设计了一个基于YOLOv5的深度神经网络来检测基于图像的货币面额。从而有助于完成真实的交易。对不同背景下的纸币图像、呈现的纸币的多个侧面、围绕杂乱物体的纸币、靠近反射表面的纸币以及模糊的纸币图像进行了鲁棒模型的训练。为开发该模型,创建了一个包含约10,000张原始图像的注释和增强数据集。采用预处理步骤将所有图像重新缩放为224 × 224,对神经网络的输入进行标准化,并将模型推广到包括单板计算机和智能手机在内的不同平台。对于完全不同的数据集,训练后的模型显示出92.71%的平均面额识别准确率。将训练好的模型分别部署在树莓派和智能手机上,并成功演示了从图像中检测货币面额的结果。该模型在不同的平台上显示了足够的性能,从而导致了基于货币识别模型的其他几个辅助应用的探索,以提高视障人士的生活水平。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
An Indian Currency Recognition Model for Assisting Visually Impaired Individuals
Visually impaired persons find it extremely difficult to perform cash transactions in outdoor environments. For assisting the visually challenged individuals, a YOLOv5 based deep neural network was designed to detect image based currency denominations. Thereby aid in completing the authentic transaction. The robust model was trained for images with currency notes in different backgrounds, multiple sides of the currency notes presented, notes around cluttered objects, notes near reflective surfaces, and blurred images of the currency notes. An annotated and augmented dataset of around 10,000 original images was created for developing the model. A pre-processing step to rescale all the images to 224 × 224 was applied to standardize the input to the neural network, and generalize the model for different platforms including single board computer and smartphones. The trained model showcased an average denomination recognition accuracy of 92.71% for an altogether different dataset. The trained model was deployed on Raspberry-Pi and Smartphone independently, and the outcome to detect the currency denomination from the image was successfully demonstrated. The model showcased adequate performance on different platforms, leading to the exploration of several other assistive applications based on the currency recognition model to improve the standard of living for visually challenged individuals.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Design 4x1 Space-Time Conjugate Two-Path Full-Rate OFDM Systems YOLO-Based Deep-Learning Gaze Estimation Technology by Combining Geometric Feature and Appearance Based Technologies for Smart Advertising Displays Graph Neural Networks for HD EMG-based Movement Intention Recognition: An Initial Investigation Bert Based Chinese Sentiment Analysis for Automatic Censoring of Dynamic Electronic Scroll An Image Feature Points Assisted Point Cloud Matching Scheme in Odometry Estimation for SLAM Systems
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1