首页 > 最新文献

International Journal of Innovative Research in Engineering最新文献

英文 中文
ML-Driven Facial Synthesis from Spoken Words Using Conditional GANs 使用条件 GAN 从口语中提取 ML 驱动的面部合成
Pub Date : 2024-01-25 DOI: 10.59256/ijire.20240501004
Vaishnavi Srivastava, Sakshi Srivastava, Sakshi Chauhan, Divyakshi Yadav
A Human Brain may translate a person's voice to its corresponding face image even if never seen before. Training adeep learning network to do the same can be used in detecting human faces based on their voice, which may be used in findinga criminal that we only have a voice recording for. The goal in this paper is to build a Conditional Generative Adversarial Network that produces face images from human speeches which can then be recognized by a face recognition model to identifythe owner of the speech. The model was trained, and the face recognition model gave an accuracy of 80.08% in training and 56.2% in testing. Compared to the basic GAN model, this model has improved the results by about 30%. Key Word: Face image synthesis, Generative adversarial network, Face Recognition
人脑可以将一个人的声音转化为相应的人脸图像,即使是从未见过的人。对深度学习网络进行同样的训练,可用于根据人的声音检测人脸,从而找到我们只有声音记录的罪犯。本文的目标是建立一个条件生成对抗网络,该网络可以从人类讲话中生成人脸图像,然后通过人脸识别模型识别出讲话的主人。模型经过训练后,人脸识别模型的训练准确率为 80.08%,测试准确率为 56.2%。与基本的 GAN 模型相比,该模型的结果提高了约 30%。关键字人脸图像合成 生成式对抗网络 人脸识别
{"title":"ML-Driven Facial Synthesis from Spoken Words Using Conditional GANs","authors":"Vaishnavi Srivastava, Sakshi Srivastava, Sakshi Chauhan, Divyakshi Yadav","doi":"10.59256/ijire.20240501004","DOIUrl":"https://doi.org/10.59256/ijire.20240501004","url":null,"abstract":"A Human Brain may translate a person's voice to its corresponding face image even if never seen before. Training adeep learning network to do the same can be used in detecting human faces based on their voice, which may be used in findinga criminal that we only have a voice recording for. The goal in this paper is to build a Conditional Generative Adversarial Network that produces face images from human speeches which can then be recognized by a face recognition model to identifythe owner of the speech. The model was trained, and the face recognition model gave an accuracy of 80.08% in training and 56.2% in testing. Compared to the basic GAN model, this model has improved the results by about 30%. Key Word: Face image synthesis, Generative adversarial network, Face Recognition","PeriodicalId":516932,"journal":{"name":"International Journal of Innovative Research in Engineering","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-01-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140495923","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BLYNK RFID and Retinal Lock Access System BLYNK 射频识别和视网膜锁门禁系统
Pub Date : 2024-01-24 DOI: 10.59256/ijire.20240501003
Yoheswari S, Adhithyaram L, Gokulesh S, Harish Raj K.B, Jivithesh Harshaa R D
The BLYNK RFID AND RETINAL LOCKACCESS SYSTEM describes a digital door lock system that uses an ESP32-CAM module, which is a budget friendly development board with a very small size camera and a micro-SD card slot. The system uses retinal recognition technology to detect the retinal of the person who wants to access the door. The AI-Thinker ESP32-CAM module takes pictures of the person and sends them to the owner via the BLYNK application installed on their mobile phone. The owner can then grant permission to access the door based on the person’s identity. When deploying your BLYNK RFID and retinal scanner project, it's important to consider scalability and maintenance. As your user base and access requirements may change over time, plan for future expansion and updates. Regularly review and update your system's firmware, libraries, and security measures to stay ahead of potential vulnerabilities and evolving best practices in access control. Monitoring and auditing your system's usage is crucial. The Blynk platform can help you gather data on access attempts and system performance, allowing you to analyze the data for any anomalies and potential security breaches. This data can be valuable for compliance, troubleshooting, and performance optimization. Key Word: retinal and RFID scanning for lock to authentic users, using an ESP32-CAM and RFID reader controlling through BLYNK.
BLYNK RFID 和视网膜门锁系统介绍了一种使用 ESP32-CAM 模块的数字门锁系统,ESP32-CAM 模块是一种经济实惠的开发板,带有一个非常小的摄像头和一个微型 SD 卡插槽。该系统使用视网膜识别技术来检测想要开门的人的视网膜。AI-Thinker ESP32-CAM 模块会拍摄该人的照片,并通过安装在手机上的 BLYNK 应用程序发送给主人。然后,主人就可以根据该人的身份授予开门权限。在部署 BLYNK RFID 和视网膜扫描仪项目时,必须考虑可扩展性和维护问题。随着时间的推移,您的用户群和访问要求可能会发生变化,因此要为未来的扩展和更新做好计划。定期检查和更新系统的固件、库和安全措施,以便及时发现潜在的漏洞和不断发展的门禁控制最佳实践。监控和审计系统的使用情况至关重要。Blynk 平台可帮助您收集有关访问尝试和系统性能的数据,使您能够对数据进行分析,查找任何异常情况和潜在的安全漏洞。这些数据对于合规性、故障排除和性能优化都很有价值。关键词:视网膜和 RFID 扫描,通过 BLYNK 控制的 ESP32-CAM 和 RFID 阅读器锁定真实用户。
{"title":"BLYNK RFID and Retinal Lock Access System","authors":"Yoheswari S, Adhithyaram L, Gokulesh S, Harish Raj K.B, Jivithesh Harshaa R D","doi":"10.59256/ijire.20240501003","DOIUrl":"https://doi.org/10.59256/ijire.20240501003","url":null,"abstract":"The BLYNK RFID AND RETINAL LOCKACCESS SYSTEM describes a digital door lock system that uses an ESP32-CAM module, which is a budget friendly development board with a very small size camera and a micro-SD card slot. The system uses retinal recognition technology to detect the retinal of the person who wants to access the door. The AI-Thinker ESP32-CAM module takes pictures of the person and sends them to the owner via the BLYNK application installed on their mobile phone. The owner can then grant permission to access the door based on the person’s identity. When deploying your BLYNK RFID and retinal scanner project, it's important to consider scalability and maintenance. As your user base and access requirements may change over time, plan for future expansion and updates. Regularly review and update your system's firmware, libraries, and security measures to stay ahead of potential vulnerabilities and evolving best practices in access control. Monitoring and auditing your system's usage is crucial. The Blynk platform can help you gather data on access attempts and system performance, allowing you to analyze the data for any anomalies and potential security breaches. This data can be valuable for compliance, troubleshooting, and performance optimization. Key Word: retinal and RFID scanning for lock to authentic users, using an ESP32-CAM and RFID reader controlling through BLYNK.","PeriodicalId":516932,"journal":{"name":"International Journal of Innovative Research in Engineering","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140497224","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Custom Voice Cloner 自定义语音克隆器
Pub Date : 2024-01-19 DOI: 10.59256/ijire.20240501002
Usharani K, Nandha kumaran H, Nikhilesh Pranav M.S, Nithish kumar K.K, Prasanna Krishna A.S
The Custom Voice Cloner is based on voice signal speech synthesizer. It is a technology that converts text into audible speech, simulating human speech characteristics like pitch and tone. It finds applications in virtual assistants, navigation systems, and accessibility tools. Building one in Python typically involves Text-to-Speech (TTS) libraries such as gTTS, pyttsx3, or platform-specific options for Windows and macOS, offering easy text-to-speech conversion.However, TTS libraries might lack customization and voice quality needed for advanced projects. For more sophisticated applications, custom voice synthesizers can be built using deep learning techniques like Tacotron and WaveNet. These models learn speech nuances for more natural output.Creating a custom voice synthesizer is challenging, requiring high-quality training data, machine learning expertise, and substantial computational resources. It goes beyond generating speech to convey emotions and nuances in pronunciation for natural and expressive voices. Key Word: Voice signal speech synthesizer,text-to-speech conversion, deep learning,TTS, gTTS, pyttsx3,etc.
自定义语音克隆器基于语音信号语音合成器。它是一种将文本转换为可听语音的技术,可模拟人的语音特征,如音高和音调。它可应用于虚拟助手、导航系统和无障碍工具。在 Python 中构建语音合成器通常需要使用文本到语音(TTS)库,如 gTTS、pyttsx3 或针对 Windows 和 macOS 平台的特定选项,这些库提供了简单的文本到语音转换功能。对于更复杂的应用,可以使用 Tacotron 和 WaveNet 等深度学习技术构建自定义语音合成器。创建定制语音合成器具有挑战性,需要高质量的训练数据、机器学习专业知识和大量计算资源。它不仅能生成语音,还能传达情感和发音上的细微差别,从而发出自然而富有表现力的声音。关键字语音信号语音合成器、文本到语音转换、深度学习、TTS、gTTS、pyttsx3 等。
{"title":"Custom Voice Cloner","authors":"Usharani K, Nandha kumaran H, Nikhilesh Pranav M.S, Nithish kumar K.K, Prasanna Krishna A.S","doi":"10.59256/ijire.20240501002","DOIUrl":"https://doi.org/10.59256/ijire.20240501002","url":null,"abstract":"The Custom Voice Cloner is based on voice signal speech synthesizer. It is a technology that converts text into audible speech, simulating human speech characteristics like pitch and tone. It finds applications in virtual assistants, navigation systems, and accessibility tools. Building one in Python typically involves Text-to-Speech (TTS) libraries such as gTTS, pyttsx3, or platform-specific options for Windows and macOS, offering easy text-to-speech conversion.However, TTS libraries might lack customization and voice quality needed for advanced projects. For more sophisticated applications, custom voice synthesizers can be built using deep learning techniques like Tacotron and WaveNet. These models learn speech nuances for more natural output.Creating a custom voice synthesizer is challenging, requiring high-quality training data, machine learning expertise, and substantial computational resources. It goes beyond generating speech to convey emotions and nuances in pronunciation for natural and expressive voices. Key Word: Voice signal speech synthesizer,text-to-speech conversion, deep learning,TTS, gTTS, pyttsx3,etc.","PeriodicalId":516932,"journal":{"name":"International Journal of Innovative Research in Engineering","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140502508","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Embedding Artificial Intelligence for Personal Voice Assistant Using NLP 利用 NLP 为个人语音助理嵌入人工智能
Pub Date : 2024-01-06 DOI: 10.59256/ijire.20240501001
Maria Sobana S, R. M, Rajkumar R K, Rajkumar M, Siddarthan S
The voice assistance is an software which is able to provide a detailed response as a voice based output according to an instruction in a prompt. To seamless integration of quick responses to queries and up-to-date weather information enhances daily routines, promoting efficiency and convenience. To achieve these capabilities, technologies like NLTK, pyttsx3, and speech recognition libraries play a pivotal role. To summarize, the convergence of these tools is gradually transforming the futuristic concept of an indispensable personal assistant into an attainable reality. AI technologies have revolutionized digital assistant interactions, but as they integrate into daily life, addressing bias, ambiguity, and ethics becomes crucial. Key Word: Integration; Convergence; Futuristic; Indispensable;
语音辅助是一种软件,能够根据提示中的指令以语音输出的形式提供详细答复。将快速回复询问和最新天气信息无缝整合在一起,可以提高日常工作的效率和便利性。为了实现这些功能,NLTK、pyttsx3 和语音识别库等技术发挥了关键作用。总之,这些工具的融合正逐步将不可或缺的个人助理这一未来概念转变为可实现的现实。人工智能技术已经彻底改变了数字助理的交互方式,但随着它们融入日常生活,解决偏见、模糊性和道德问题变得至关重要。关键字整合;融合;未来主义;不可或缺;
{"title":"Embedding Artificial Intelligence for Personal Voice Assistant Using NLP","authors":"Maria Sobana S, R. M, Rajkumar R K, Rajkumar M, Siddarthan S","doi":"10.59256/ijire.20240501001","DOIUrl":"https://doi.org/10.59256/ijire.20240501001","url":null,"abstract":"The voice assistance is an software which is able to provide a detailed response as a voice based output according to an instruction in a prompt. To seamless integration of quick responses to queries and up-to-date weather information enhances daily routines, promoting efficiency and convenience. To achieve these capabilities, technologies like NLTK, pyttsx3, and speech recognition libraries play a pivotal role. To summarize, the convergence of these tools is gradually transforming the futuristic concept of an indispensable personal assistant into an attainable reality. AI technologies have revolutionized digital assistant interactions, but as they integrate into daily life, addressing bias, ambiguity, and ethics becomes crucial. Key Word: Integration; Convergence; Futuristic; Indispensable;","PeriodicalId":516932,"journal":{"name":"International Journal of Innovative Research in Engineering","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-01-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140513048","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Innovative Research in Engineering
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1