首页 > 最新文献

2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP)最新文献

英文 中文
[ICCCSP 2020 Front cover] [ICCCSP 2020封面]
Pub Date : 2020-09-28 DOI: 10.1109/icccsp49186.2020.9315249
{"title":"[ICCCSP 2020 Front cover]","authors":"","doi":"10.1109/icccsp49186.2020.9315249","DOIUrl":"https://doi.org/10.1109/icccsp49186.2020.9315249","url":null,"abstract":"","PeriodicalId":310458,"journal":{"name":"2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP)","volume":"24 18","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120859227","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving Facial Recognition of FaceNet in a small dataset using DeepFakes 使用DeepFakes在小数据集中改进FaceNet的面部识别
Pub Date : 2020-09-28 DOI: 10.1109/ICCCSP49186.2020.9315253
Ansh Abhay Balde, Abhay Jain, D. Patra
Facial recognition is affected by many factors such as low-resolution images, availability of datasets, illumination, pose invariance, ageing, expression, etc. With the increasing availability of powerful GPUs, we are using massive datasets to facilitate better accuracy. Among different datasets available to perform training of any facial recognition algorithm, very few of them can be a run on a low configuration system. Still, the same can’t be used to create satisfactory FaceSwap results because of anomalies in them. A small dataset of 20 identities has been created on which the results of this paper are observed. This paper introduces the usage of DeepFakes algorithm to improve the performance of FaceNet with SqueezeNet architecture and softmax loss function. It is expected that more data leads to better performance. Then the FaceSwap variation of DeepFakes is used to swap identities and create fake images for a given identity. Then, FaceNet is used to identify faces on the newly formed dataset 250RF using fake images and the original dataset 200R. This method achieves satisfactory results on training and testing accuracy in comparison, thereby creating prospects on such a method.
人脸识别受到许多因素的影响,如低分辨率图像、数据集的可用性、光照、姿态不变性、老化、表情等。随着功能强大的gpu的可用性越来越高,我们正在使用大量数据集来提高准确性。在可用于任何面部识别算法训练的不同数据集中,很少有数据集可以在低配置系统上运行。然而,同样的方法不能产生令人满意的FaceSwap结果,因为其中存在异常。创建了一个包含20个恒等式的小数据集,在此基础上观察了本文的结果。本文介绍了使用DeepFakes算法来提高FaceNet的性能,该算法采用了SqueezeNet架构和softmax损失函数。数据越多,性能越好。然后使用DeepFakes的FaceSwap变体来交换身份并为给定身份创建假图像。然后,使用FaceNet在新形成的数据集250RF上使用假图像和原始数据集200R进行人脸识别。通过对比,该方法在训练精度和测试精度上都取得了令人满意的结果,从而开创了该方法的发展前景。
{"title":"Improving Facial Recognition of FaceNet in a small dataset using DeepFakes","authors":"Ansh Abhay Balde, Abhay Jain, D. Patra","doi":"10.1109/ICCCSP49186.2020.9315253","DOIUrl":"https://doi.org/10.1109/ICCCSP49186.2020.9315253","url":null,"abstract":"Facial recognition is affected by many factors such as low-resolution images, availability of datasets, illumination, pose invariance, ageing, expression, etc. With the increasing availability of powerful GPUs, we are using massive datasets to facilitate better accuracy. Among different datasets available to perform training of any facial recognition algorithm, very few of them can be a run on a low configuration system. Still, the same can’t be used to create satisfactory FaceSwap results because of anomalies in them. A small dataset of 20 identities has been created on which the results of this paper are observed. This paper introduces the usage of DeepFakes algorithm to improve the performance of FaceNet with SqueezeNet architecture and softmax loss function. It is expected that more data leads to better performance. Then the FaceSwap variation of DeepFakes is used to swap identities and create fake images for a given identity. Then, FaceNet is used to identify faces on the newly formed dataset 250RF using fake images and the original dataset 200R. This method achieves satisfactory results on training and testing accuracy in comparison, thereby creating prospects on such a method.","PeriodicalId":310458,"journal":{"name":"2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114439302","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Detecting Trojan Attacks on Deep Neural Networks 基于深度神经网络的木马攻击检测
Pub Date : 2020-09-28 DOI: 10.1109/ICCCSP49186.2020.9315256
J. Singh, V. Sharmila
Machine learning and Artificial Intelligent techniques are the most used techniques. It gives opportunity to online sharing market where sharing and adopting model is being popular. It gives attackers many new opportunities. Deep neural network is the most used approached for artificial techniques. In this paper we are presenting a Proof of Concept method to detect Trojan attacks on the Deep Neural Network. Deploying trojan models can be dangerous in normal human lives (Application like Automated vehicle). First inverse the neuron network to create general trojan triggers, and then retrain the model with external datasets to inject Trojan trigger to the model. The malicious behaviors are only activated with the trojan trigger Input. In attack, original datasets are not required to train the model. In practice, usually datasets are not shared due to privacy or copyright concerns. We use five different applications to demonstrate the attack, and perform an analysis on the factors that affect the attack. The behavior of a trojan modification can be triggered without affecting the test accuracy for normal input datasets. After generating the trojan trigger and performing an attack. It's applying SHAP as defense against such attacks. SHAP is known for its unique explanation for model predictions.
机器学习和人工智能技术是最常用的技术。这给正在流行分享和采用模式的在线分享市场带来了机会。它给攻击者提供了许多新的机会。深度神经网络是人工技术中最常用的方法。在本文中,我们提出了一种检测深度神经网络上木马攻击的概念验证方法。部署木马模型在正常的人类生活中可能是危险的(应用程序如自动车辆)。首先对神经元网络进行逆运算,生成通用的木马触发器,然后利用外部数据集对模型进行再训练,将木马触发器注入模型。恶意行为只有在输入木马触发器时才会被激活。在攻击中,不需要原始数据集来训练模型。在实践中,由于隐私或版权问题,通常不共享数据集。我们使用五个不同的应用程序来演示攻击,并对影响攻击的因素进行分析。可以触发木马修改的行为,而不会影响正常输入数据集的测试准确性。生成木马触发器并执行攻击后。它正在应用SHAP作为对此类攻击的防御。SHAP以其对模型预测的独特解释而闻名。
{"title":"Detecting Trojan Attacks on Deep Neural Networks","authors":"J. Singh, V. Sharmila","doi":"10.1109/ICCCSP49186.2020.9315256","DOIUrl":"https://doi.org/10.1109/ICCCSP49186.2020.9315256","url":null,"abstract":"Machine learning and Artificial Intelligent techniques are the most used techniques. It gives opportunity to online sharing market where sharing and adopting model is being popular. It gives attackers many new opportunities. Deep neural network is the most used approached for artificial techniques. In this paper we are presenting a Proof of Concept method to detect Trojan attacks on the Deep Neural Network. Deploying trojan models can be dangerous in normal human lives (Application like Automated vehicle). First inverse the neuron network to create general trojan triggers, and then retrain the model with external datasets to inject Trojan trigger to the model. The malicious behaviors are only activated with the trojan trigger Input. In attack, original datasets are not required to train the model. In practice, usually datasets are not shared due to privacy or copyright concerns. We use five different applications to demonstrate the attack, and perform an analysis on the factors that affect the attack. The behavior of a trojan modification can be triggered without affecting the test accuracy for normal input datasets. After generating the trojan trigger and performing an attack. It's applying SHAP as defense against such attacks. SHAP is known for its unique explanation for model predictions.","PeriodicalId":310458,"journal":{"name":"2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128080945","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Exploration of Big Data Analytics in Healthcare Analytics 医疗保健分析中的大数据分析探索
Pub Date : 2020-09-28 DOI: 10.1109/ICCCSP49186.2020.9315192
R. Katarya, Sajal Jain
The data is increasing in every field of business finance etc. But the health sector field is not yet fully explored and there is a lot of scope of advancement there. In the past few years, the health sector had gone through a lot of changes, there is a high increase in the number of doctors, patients and diseases. The data that once was in the size of terabytes now reached zettabytes and still growing exponentially. As the data is collected from various sources, therefore, it can be both structured and unstructured. So a lot of time and money is wasted on storing and analyzing it. That’s why it is very difficult and complex to analyze it using traditional approaches. So as the time is passing by it is clear that the use of big data tools and techniques will become a necessity in every field. Various scientist and organizations are trying to use these techniques for their financial advantage. Big data analytics will help us to analyze and find the pattern between the data sets by the help of which we can improve the state of the current healthcare system.
数据在商业、金融等各个领域都在增长。但卫生领域尚未得到充分探索,还有很大的发展空间。在过去几年中,卫生部门经历了许多变化,医生、病人和疾病的数量都大幅增加。曾经以太字节为单位的数据现在达到了泽字节,而且还在呈指数级增长。由于数据是从各种来源收集的,因此它可以是结构化的,也可以是非结构化的。因此,大量的时间和金钱被浪费在存储和分析数据上。这就是为什么用传统的方法来分析它是非常困难和复杂的。因此,随着时间的流逝,大数据工具和技术的使用显然将成为每个领域的必需品。许多科学家和组织正试图利用这些技术来获取经济利益。大数据分析将帮助我们分析和发现数据集之间的模式,通过这种模式,我们可以改善当前医疗系统的状态。
{"title":"Exploration of Big Data Analytics in Healthcare Analytics","authors":"R. Katarya, Sajal Jain","doi":"10.1109/ICCCSP49186.2020.9315192","DOIUrl":"https://doi.org/10.1109/ICCCSP49186.2020.9315192","url":null,"abstract":"The data is increasing in every field of business finance etc. But the health sector field is not yet fully explored and there is a lot of scope of advancement there. In the past few years, the health sector had gone through a lot of changes, there is a high increase in the number of doctors, patients and diseases. The data that once was in the size of terabytes now reached zettabytes and still growing exponentially. As the data is collected from various sources, therefore, it can be both structured and unstructured. So a lot of time and money is wasted on storing and analyzing it. That’s why it is very difficult and complex to analyze it using traditional approaches. So as the time is passing by it is clear that the use of big data tools and techniques will become a necessity in every field. Various scientist and organizations are trying to use these techniques for their financial advantage. Big data analytics will help us to analyze and find the pattern between the data sets by the help of which we can improve the state of the current healthcare system.","PeriodicalId":310458,"journal":{"name":"2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116288342","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Secured Automated Complaint Generation System for Organizations 为机构提供安全的自动投诉生成系统
Pub Date : 2020-09-28 DOI: 10.1109/ICCCSP49186.2020.9315243
K. Vidya, Lavanya Amalbabu, K.S Sowndharya, S. balaji
In our society, complaint systems are all done manually by humans and there are no automated system wherein the complaints are identified and sent to the respective authority all by itself. A complaint about our day to day activities based on the images can be identified and given as a report. This reduces lot of manual time and may speedup the remedial process. In this system, the users need to upload an image which will be analyzed and classified based on remedial departments. Thus the amount of manual work both on the user and the organization is minimized by feasibly producing an automated system that can generate a report about the problem without human intervention during the process. Based on the institutional requirements, the complaints are classified as website-based and object-based using an image classification system. The web-related complaints are handled by optical character recognition and the object-based complaints are handled by object detection and data mining techniques. The images are trained and tested through various classification system and their performances are compared. The user is also provided with a feature of adding location along with the complaint image which makes less complication in finding the place of fault and based on which a report is generated and forwarded to corresponding department. Thus, this paper aims in proposing an automated complaint generation and reporting system for Institutions by classifying user input images using image processing and data mining techniques.
在我们的社会中,投诉系统都是由人类手动完成的,没有自动识别投诉并将其发送给相应当局的自动系统。对我们日常活动的投诉可以根据图像进行识别并作为报告给出。这减少了大量的人工时间,并可能加快补救过程。在本系统中,用户需要上传一张图片,然后根据补救部门对图片进行分析和分类。因此,通过可行地生成一个自动化系统,用户和组织的手工工作量都被最小化,该系统可以在过程中生成关于问题的报告,而无需人工干预。根据机构要求,使用图像分类系统将投诉分为基于网站和基于对象的投诉。网络投诉采用光学字符识别处理,基于对象的投诉采用对象检测和数据挖掘技术处理。通过各种分类系统对图像进行训练和测试,并对其性能进行比较。用户还具有随投诉图像添加位置的功能,减少了查找故障地点的复杂性,并根据该位置生成报告并转发给相应部门。因此,本文旨在通过使用图像处理和数据挖掘技术对用户输入的图像进行分类,为机构提出一个自动投诉生成和报告系统。
{"title":"Secured Automated Complaint Generation System for Organizations","authors":"K. Vidya, Lavanya Amalbabu, K.S Sowndharya, S. balaji","doi":"10.1109/ICCCSP49186.2020.9315243","DOIUrl":"https://doi.org/10.1109/ICCCSP49186.2020.9315243","url":null,"abstract":"In our society, complaint systems are all done manually by humans and there are no automated system wherein the complaints are identified and sent to the respective authority all by itself. A complaint about our day to day activities based on the images can be identified and given as a report. This reduces lot of manual time and may speedup the remedial process. In this system, the users need to upload an image which will be analyzed and classified based on remedial departments. Thus the amount of manual work both on the user and the organization is minimized by feasibly producing an automated system that can generate a report about the problem without human intervention during the process. Based on the institutional requirements, the complaints are classified as website-based and object-based using an image classification system. The web-related complaints are handled by optical character recognition and the object-based complaints are handled by object detection and data mining techniques. The images are trained and tested through various classification system and their performances are compared. The user is also provided with a feature of adding location along with the complaint image which makes less complication in finding the place of fault and based on which a report is generated and forwarded to corresponding department. Thus, this paper aims in proposing an automated complaint generation and reporting system for Institutions by classifying user input images using image processing and data mining techniques.","PeriodicalId":310458,"journal":{"name":"2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129783203","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Speech-Based Virtual Travel Assistant For Visually Impaired 基于语音的视障虚拟旅行助手
Pub Date : 2020-09-28 DOI: 10.1109/ICCCSP49186.2020.9315217
N. Sripriya, S. Poornima, S. Mohanavalli, R. Bhaiya, V. Nikita
Visually-impaired people often feel handicapped and find it difficult to explore the world around them freely without any kind of help and assistance. This lack of independence restricts them from feeling confident about themselves. Furthermore, all virtual travel assistants currently existing in the market, are chatbots associated with textual interaction. While this may be useful more often than not, in certain situations, users may find textual interaction uncomfortable and prefer a conversational interface over text. An important issue of concern in the tourism industry is the monetary expense invested in travel guides by tourists. A travel assistant capable of acting as a virtual guide can help reduce this investment by a large factor. Hence, it is proposed to develop a speech-based travel bot capable of acting as a virtual tour guide. The bot will play the role of a tour guide by suggesting places and giving information about the place such as opening hours, rating, address to aid the user in knowing more about the place by interacting with the user and providing relevant information. The proposed system can also find great use in situations where a person may feel uncomfortable to text such as when driving. This speech-based virtual travel assistant/bot is implemented using speech recognition, speech synthesis, and natural language techniques to gather the user’s preference and provide the intended output. The developed bot has proved to be efficient for searching for different kinds of places and interacts well with the user and helps to find further details about a specific place as per the user’s queries. The intents for most of the queries are correctly recognized, which helps in efficient dialogue management.
视障人士经常觉得自己是残疾人,在没有任何帮助和帮助的情况下,很难自由地探索周围的世界。这种独立性的缺乏限制了他们对自己的信心。此外,目前市场上所有的虚拟旅行助手都是与文本交互相关的聊天机器人。虽然这通常是有用的,但在某些情况下,用户可能会发现文本交互不舒服,并且更喜欢会话界面而不是文本。旅游业关注的一个重要问题是游客在导游上投入的金钱费用。一个能够充当虚拟导游的旅行助理可以在很大程度上减少这种投资。因此,建议开发一种能够充当虚拟导游的基于语音的旅行机器人。该机器人将扮演导游的角色,通过推荐地点并提供有关地点的信息,如开放时间、评级、地址等,帮助用户通过与用户互动并提供相关信息来更多地了解这个地方。该系统还可以在人们不太愿意发短信的情况下发挥很大作用,比如开车的时候。这个基于语音的虚拟旅行助手/机器人使用语音识别、语音合成和自然语言技术来收集用户的偏好并提供预期的输出。开发的机器人已被证明可以高效地搜索不同类型的地点,并与用户进行良好的交互,并有助于根据用户的查询找到有关特定地点的进一步细节。可以正确识别大多数查询的意图,这有助于有效地管理对话。
{"title":"Speech-Based Virtual Travel Assistant For Visually Impaired","authors":"N. Sripriya, S. Poornima, S. Mohanavalli, R. Bhaiya, V. Nikita","doi":"10.1109/ICCCSP49186.2020.9315217","DOIUrl":"https://doi.org/10.1109/ICCCSP49186.2020.9315217","url":null,"abstract":"Visually-impaired people often feel handicapped and find it difficult to explore the world around them freely without any kind of help and assistance. This lack of independence restricts them from feeling confident about themselves. Furthermore, all virtual travel assistants currently existing in the market, are chatbots associated with textual interaction. While this may be useful more often than not, in certain situations, users may find textual interaction uncomfortable and prefer a conversational interface over text. An important issue of concern in the tourism industry is the monetary expense invested in travel guides by tourists. A travel assistant capable of acting as a virtual guide can help reduce this investment by a large factor. Hence, it is proposed to develop a speech-based travel bot capable of acting as a virtual tour guide. The bot will play the role of a tour guide by suggesting places and giving information about the place such as opening hours, rating, address to aid the user in knowing more about the place by interacting with the user and providing relevant information. The proposed system can also find great use in situations where a person may feel uncomfortable to text such as when driving. This speech-based virtual travel assistant/bot is implemented using speech recognition, speech synthesis, and natural language techniques to gather the user’s preference and provide the intended output. The developed bot has proved to be efficient for searching for different kinds of places and interacts well with the user and helps to find further details about a specific place as per the user’s queries. The intents for most of the queries are correctly recognized, which helps in efficient dialogue management.","PeriodicalId":310458,"journal":{"name":"2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116571249","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Data Compression Algorithm for Audio and Image using Feature Extraction 基于特征提取的音频和图像数据压缩算法
Pub Date : 2020-09-28 DOI: 10.1109/ICCCSP49186.2020.9315248
Mohammad Sheraj, Ashish Chopra
We aim to achieve the highest data compression ratio in a lossy scenario while still maintaining the original image or audio files characteristics and resolution/bitrate. For this we would run feature extraction on chunks of the data and store them in a database with a specific hash as a key. This hash will be stored in the file and the full data later reconstructed from the database. The database will be created by training on a vast range of data and storing only the most common chunks encountered by hash. The compression ratio achieved for image it is 0.01 over standard raw input data.
我们的目标是在有损的情况下实现最高的数据压缩比,同时仍然保持原始图像或音频文件的特征和分辨率/比特率。为此,我们将对数据块进行特征提取,并将它们存储在数据库中,并使用特定的哈希作为键。该散列将存储在文件中,然后从数据库中重建完整的数据。数据库将通过对大量数据进行训练并仅存储哈希遇到的最常见块来创建。与标准原始输入数据相比,图像的压缩比为0.01。
{"title":"Data Compression Algorithm for Audio and Image using Feature Extraction","authors":"Mohammad Sheraj, Ashish Chopra","doi":"10.1109/ICCCSP49186.2020.9315248","DOIUrl":"https://doi.org/10.1109/ICCCSP49186.2020.9315248","url":null,"abstract":"We aim to achieve the highest data compression ratio in a lossy scenario while still maintaining the original image or audio files characteristics and resolution/bitrate. For this we would run feature extraction on chunks of the data and store them in a database with a specific hash as a key. This hash will be stored in the file and the full data later reconstructed from the database. The database will be created by training on a vast range of data and storing only the most common chunks encountered by hash. The compression ratio achieved for image it is 0.01 over standard raw input data.","PeriodicalId":310458,"journal":{"name":"2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126751837","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Design of Imitative Control Modalities for a 3 Degree of Freedom Robotic Arm 三自由度机械臂的仿真控制模式设计
Pub Date : 2020-09-28 DOI: 10.1109/ICCCSP49186.2020.9315273
H. Gokul, S. V. Kanna, H. Akshay Kumar, Vignesh Ravikumar
This research article focuses on enabling real-time Human-Robot Interaction (HRI) with two input control modalities designed to control a robotic arm. The robotic arm comprises of 3 Degrees of Freedom (DOF) and is controlled using the following 2 strategies: Inertial Sensor-glove and Image-based visual servoing. The Mathematical model of the given 3 DOF robotic arm was derived and the two proposed methods were explicitly designed for the same. The inertial sensor-glove, worn by the user, is embedded with inertial sensors that provide data to track the motion of the user's arm. The Image-based visual servoing modality is a mono-vision based approach which tracks the motion of a target object held by the user using a camera and operates similar to a mouse-pointer but in 3 dimensions. Fitts’s Targeting tasks were performed to analyse the performance of these Human-Robot Interactive input modalities.
本文的研究重点是实现实时人机交互(HRI)与两种输入控制模式的设计,以控制机械臂。机械臂由3个自由度组成,采用以下两种策略进行控制:惯性传感器手套和基于图像的视觉伺服。推导了给定三自由度机械臂的数学模型,并针对该模型明确设计了两种方法。用户佩戴的惯性传感器手套嵌入了惯性传感器,可以提供数据来跟踪用户手臂的运动。基于图像的视觉伺服模式是一种基于单视觉的方法,它跟踪用户使用相机持有的目标物体的运动,操作类似于鼠标指针,但在三维空间中。执行Fitts的目标任务来分析这些人机交互输入模式的性能。
{"title":"Design of Imitative Control Modalities for a 3 Degree of Freedom Robotic Arm","authors":"H. Gokul, S. V. Kanna, H. Akshay Kumar, Vignesh Ravikumar","doi":"10.1109/ICCCSP49186.2020.9315273","DOIUrl":"https://doi.org/10.1109/ICCCSP49186.2020.9315273","url":null,"abstract":"This research article focuses on enabling real-time Human-Robot Interaction (HRI) with two input control modalities designed to control a robotic arm. The robotic arm comprises of 3 Degrees of Freedom (DOF) and is controlled using the following 2 strategies: Inertial Sensor-glove and Image-based visual servoing. The Mathematical model of the given 3 DOF robotic arm was derived and the two proposed methods were explicitly designed for the same. The inertial sensor-glove, worn by the user, is embedded with inertial sensors that provide data to track the motion of the user's arm. The Image-based visual servoing modality is a mono-vision based approach which tracks the motion of a target object held by the user using a camera and operates similar to a mouse-pointer but in 3 dimensions. Fitts’s Targeting tasks were performed to analyse the performance of these Human-Robot Interactive input modalities.","PeriodicalId":310458,"journal":{"name":"2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126840952","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Detection of sharp objects using deep neural network based object detection algorithm 锐利物体检测采用基于深度神经网络的物体检测算法
Pub Date : 2020-09-28 DOI: 10.1109/ICCCSP49186.2020.9315272
R. Kayalvizhi, S. Malarvizhi, S. Choudhury, A. Topkar, P. Vijayakumar
Deep learning algorithms have the ability to learn complex functions and provide state-of-the-art results for com-puter vision problems. In recent times, these algorithms far exceeded the existing computer vision based techniques for object detection in X-ray imaging systems. So far, in literature single class of object namely gun and its parts were considered for detection using the SIXray10 database. We propose deep learning-based solution for the detection of sharp objects namely knife, scissors, wrench, pliers in the SIXray 10database. We propose two models namely model A and model B using a common object detection algorithm- YOLOv3 (You Only Look Once) with InceptionV3 and ResNet-50. YOLO is a deep neural network based object detection algorithm that performs the task in one-shot which allows real time inference in video of 15-30 fps. The model is FCN (Fully Convolutional Network) as has the capacity to perform both regression and classification by sharing weights for both the tasks. The network predicts a rectangular box called bounding box around the predicted object of interest along with the associated class. We analyze the performance of both model in terms of mAP. We achieve mean accuracy of 59.95% for model-A and 63.35% for Model-B. The most daunting part of the project is the low ratio of harmful to nonharmful items. By performing rigorous experiments we came up with the best set of possible results which uses varied pretrained neural networks for feature extraction in tandem with YOLO model for object detection. We endeavor to improve on these existing results so as these systems can be successfully deployed in airports to minimize human error and improve security in such environments.
深度学习算法具有学习复杂函数的能力,并为计算机视觉问题提供最先进的结果。近年来,这些算法远远超过了现有的基于计算机视觉的x射线成像系统中的目标检测技术。到目前为止,文献中只考虑使用SIXray10数据库检测一类物体,即枪支及其部件。我们提出了基于深度学习的解决方案来检测尖锐物体,即SIXray 10数据库中的刀、剪刀、扳手、钳子。我们提出了两个模型,即模型A和模型B,使用通用的目标检测算法- YOLOv3(你只看一次)与InceptionV3和ResNet-50。YOLO是一种基于深度神经网络的目标检测算法,它可以在15-30 fps的视频中进行实时推理。该模型是FCN(全卷积网络),它具有通过共享两个任务的权重来执行回归和分类的能力。该网络在预测的感兴趣的对象周围预测一个称为边界框的矩形框以及相关的类。我们从mAP的角度分析了两种模型的性能。模型a和模型b的平均准确率分别为59.95%和63.35%。该项目最令人生畏的部分是有害物品与无害物品的低比例。通过严格的实验,我们提出了一组最好的可能结果,使用各种预训练的神经网络进行特征提取,并结合YOLO模型进行目标检测。我们努力改进这些现有的结果,使这些系统能够成功地部署在机场,以尽量减少人为错误,提高这种环境下的安全性。
{"title":"Detection of sharp objects using deep neural network based object detection algorithm","authors":"R. Kayalvizhi, S. Malarvizhi, S. Choudhury, A. Topkar, P. Vijayakumar","doi":"10.1109/ICCCSP49186.2020.9315272","DOIUrl":"https://doi.org/10.1109/ICCCSP49186.2020.9315272","url":null,"abstract":"Deep learning algorithms have the ability to learn complex functions and provide state-of-the-art results for com-puter vision problems. In recent times, these algorithms far exceeded the existing computer vision based techniques for object detection in X-ray imaging systems. So far, in literature single class of object namely gun and its parts were considered for detection using the SIXray10 database. We propose deep learning-based solution for the detection of sharp objects namely knife, scissors, wrench, pliers in the SIXray 10database. We propose two models namely model A and model B using a common object detection algorithm- YOLOv3 (You Only Look Once) with InceptionV3 and ResNet-50. YOLO is a deep neural network based object detection algorithm that performs the task in one-shot which allows real time inference in video of 15-30 fps. The model is FCN (Fully Convolutional Network) as has the capacity to perform both regression and classification by sharing weights for both the tasks. The network predicts a rectangular box called bounding box around the predicted object of interest along with the associated class. We analyze the performance of both model in terms of mAP. We achieve mean accuracy of 59.95% for model-A and 63.35% for Model-B. The most daunting part of the project is the low ratio of harmful to nonharmful items. By performing rigorous experiments we came up with the best set of possible results which uses varied pretrained neural networks for feature extraction in tandem with YOLO model for object detection. We endeavor to improve on these existing results so as these systems can be successfully deployed in airports to minimize human error and improve security in such environments.","PeriodicalId":310458,"journal":{"name":"2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126411474","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An Exploratory of Hybrid Techniques on Deep Learning for Image Classification 图像分类中深度学习混合技术的探索
Pub Date : 2020-09-28 DOI: 10.1109/ICCCSP49186.2020.9315270
M. Suganthi, J. Sathiaseelan
Image Mining is an extension of data mining, which is concerned with extracting beneficial information from image data. Images are classified based on texture, size, color and morphology. Neural Networks, ImageNet, VGG16, AlexNet are renowned image recognition techniques used to identify various agriculture, medical, aerial images and so on. Convolution neural network (CNN) is a Machine learning method used to classify the images which are popularly known for robust feature extraction and information mining. A comparative study of seven CNN based hybrid image classification techniques namely CNN-ELM, CNN-KNN, CNN-GA, MLP-CNN, CNN-SVM, CNN-RNN, CNN-LSTM has been done to determine their accuracy.
图像挖掘是数据挖掘的一种扩展,涉及从图像数据中提取有益信息。图像根据纹理、大小、颜色和形态进行分类。神经网络、ImageNet、VGG16、AlexNet是著名的图像识别技术,用于识别各种农业、医疗、航空图像等。卷积神经网络(CNN)是一种用于图像分类的机器学习方法,以鲁棒特征提取和信息挖掘而闻名。对7种基于CNN的混合图像分类技术CNN- elm、CNN- knn、CNN- ga、MLP-CNN、CNN- svm、CNN- rnn、CNN- lstm进行了对比研究,确定了它们的准确率。
{"title":"An Exploratory of Hybrid Techniques on Deep Learning for Image Classification","authors":"M. Suganthi, J. Sathiaseelan","doi":"10.1109/ICCCSP49186.2020.9315270","DOIUrl":"https://doi.org/10.1109/ICCCSP49186.2020.9315270","url":null,"abstract":"Image Mining is an extension of data mining, which is concerned with extracting beneficial information from image data. Images are classified based on texture, size, color and morphology. Neural Networks, ImageNet, VGG16, AlexNet are renowned image recognition techniques used to identify various agriculture, medical, aerial images and so on. Convolution neural network (CNN) is a Machine learning method used to classify the images which are popularly known for robust feature extraction and information mining. A comparative study of seven CNN based hybrid image classification techniques namely CNN-ELM, CNN-KNN, CNN-GA, MLP-CNN, CNN-SVM, CNN-RNN, CNN-LSTM has been done to determine their accuracy.","PeriodicalId":310458,"journal":{"name":"2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115353855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1