Human Activity Recognition in Smart Home using Deep Learning Techniques

Ranjit P. Kolkar, V. Geetha
{"title":"Human Activity Recognition in Smart Home using Deep Learning Techniques","authors":"Ranjit P. Kolkar, V. Geetha","doi":"10.1109/ICTS52701.2021.9609044","DOIUrl":null,"url":null,"abstract":"To understand the human activities and anticipate his intentions Human Activity Recognition(HAR) research is rapidly developing in tandem with the widespread availability of sensors. Various applications like elderly care and health monitoring systems in smart homes use smartphones and wearable devices. This paper proposes an effective HAR framework that uses deep learning methodology like Convolution Neural Networks(CNN), variations of LSTM(Long Short term Memory) and Gated Recurrent Units(GRU) Networks to recognize the activities based on smartphone sensors. The hybrid use of CNN-LSTM eliminates the handcrafted feature engineering and uses spatial and temporal data deep. The experiments are carried on UCI HAR and WISDM data sets, and the comparison results are obtained. The result shows a better 96.83 % and 98.00% for the UCI-HAR and WISDM datasets, respectively.","PeriodicalId":6738,"journal":{"name":"2021 13th International Conference on Information & Communication Technology and System (ICTS)","volume":"2 1","pages":"230-234"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 13th International Conference on Information & Communication Technology and System (ICTS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICTS52701.2021.9609044","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 6

Abstract

To understand the human activities and anticipate his intentions Human Activity Recognition(HAR) research is rapidly developing in tandem with the widespread availability of sensors. Various applications like elderly care and health monitoring systems in smart homes use smartphones and wearable devices. This paper proposes an effective HAR framework that uses deep learning methodology like Convolution Neural Networks(CNN), variations of LSTM(Long Short term Memory) and Gated Recurrent Units(GRU) Networks to recognize the activities based on smartphone sensors. The hybrid use of CNN-LSTM eliminates the handcrafted feature engineering and uses spatial and temporal data deep. The experiments are carried on UCI HAR and WISDM data sets, and the comparison results are obtained. The result shows a better 96.83 % and 98.00% for the UCI-HAR and WISDM datasets, respectively.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
使用深度学习技术的智能家居中的人类活动识别
为了了解人类活动并预测其意图,随着传感器的广泛应用,人类活动识别(HAR)研究正在迅速发展。智能家居中的老年人护理和健康监测系统等各种应用都使用智能手机和可穿戴设备。本文提出了一个有效的HAR框架,该框架使用深度学习方法,如卷积神经网络(CNN)、LSTM(长短期记忆)的变体和门控循环单元(GRU)网络来识别基于智能手机传感器的活动。CNN-LSTM的混合使用消除了手工特征工程,并深入利用了时空数据。在UCI HAR和WISDM数据集上进行了实验,得到了比较结果。结果表明,UCI-HAR和WISDM数据集的识别率分别为96.83%和98.00%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
[Copyright notice] Outlier Detection and Decision Tree for Wireless Sensor Network Fault Diagnosis Graph Algorithm for Anomaly Prediction in East Java Student Admission System FarmEasy: An Intelligent Platform to Empower Crops Prediction and Crops Marketing Hiding Messages in Audio using Modulus Operation and Simple Partition
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1