Video-Based Lifting Action Recognition Using Rank-Altered Kinematic Feature Pairs.

IF 2.9 3区 心理学 Q1 BEHAVIORAL SCIENCES Human Factors Pub Date : 2024-12-26 DOI:10.1177/00187208241309748
SeHee Jung, Bingyi Su, Lu Lu, Liwei Qing, Xu Xu
{"title":"Video-Based Lifting Action Recognition Using Rank-Altered Kinematic Feature Pairs.","authors":"SeHee Jung, Bingyi Su, Lu Lu, Liwei Qing, Xu Xu","doi":"10.1177/00187208241309748","DOIUrl":null,"url":null,"abstract":"<p><strong>Objective: </strong>To identify lifting actions and count the number of lifts performed in videos based on robust class prediction and a streamlined process for reliable real-time monitoring of lifting tasks.</p><p><strong>Background: </strong>Traditional methods for recognizing lifting actions often rely on deep learning classifiers applied to human motion data collected from wearable sensors. Despite their high performance, these methods can be difficult to implement on systems with limited hardware resources.</p><p><strong>Method: </strong>The proposed method follows a five-stage process: (1) BlazePose, a real-time pose estimation model, detects key joints of the human body. (2) These joints are preprocessed by smoothing, centering, and scaling techniques. (3) Kinematic features are extracted from the preprocessed joints. (4) Video frames are classified as lifting or nonlifting using rank-altered kinematic feature pairs. (5) A lifting counting algorithm counts the number of lifts based on the class predictions.</p><p><strong>Results: </strong>Nine rank-altered kinematic feature pairs are identified as key pairs. These pairs were used to construct an ensemble classifier, which achieved 0.89 or above in classification metrics, including accuracy, precision, recall, and F1 score. This classifier showed an accuracy of 0.90 in lifting counting and a latency of 0.06 ms, which is at least 12.5 times faster than baseline classifiers.</p><p><strong>Conclusion: </strong>This study demonstrates that computer vision-based kinematic features could be adopted to effectively and efficiently recognize lifting actions.</p><p><strong>Application: </strong>The proposed method could be deployed on various platforms, including mobile devices and embedded systems, to monitor lifting tasks in real-time for the proactive prevention of work-related low-back injuries.</p>","PeriodicalId":56333,"journal":{"name":"Human Factors","volume":" ","pages":"187208241309748"},"PeriodicalIF":2.9000,"publicationDate":"2024-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Human Factors","FirstCategoryId":"102","ListUrlMain":"https://doi.org/10.1177/00187208241309748","RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"BEHAVIORAL SCIENCES","Score":null,"Total":0}
引用次数: 0

Abstract

Objective: To identify lifting actions and count the number of lifts performed in videos based on robust class prediction and a streamlined process for reliable real-time monitoring of lifting tasks.

Background: Traditional methods for recognizing lifting actions often rely on deep learning classifiers applied to human motion data collected from wearable sensors. Despite their high performance, these methods can be difficult to implement on systems with limited hardware resources.

Method: The proposed method follows a five-stage process: (1) BlazePose, a real-time pose estimation model, detects key joints of the human body. (2) These joints are preprocessed by smoothing, centering, and scaling techniques. (3) Kinematic features are extracted from the preprocessed joints. (4) Video frames are classified as lifting or nonlifting using rank-altered kinematic feature pairs. (5) A lifting counting algorithm counts the number of lifts based on the class predictions.

Results: Nine rank-altered kinematic feature pairs are identified as key pairs. These pairs were used to construct an ensemble classifier, which achieved 0.89 or above in classification metrics, including accuracy, precision, recall, and F1 score. This classifier showed an accuracy of 0.90 in lifting counting and a latency of 0.06 ms, which is at least 12.5 times faster than baseline classifiers.

Conclusion: This study demonstrates that computer vision-based kinematic features could be adopted to effectively and efficiently recognize lifting actions.

Application: The proposed method could be deployed on various platforms, including mobile devices and embedded systems, to monitor lifting tasks in real-time for the proactive prevention of work-related low-back injuries.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于秩变运动特征对的视频举重动作识别。
目的:基于稳健的类别预测和可靠的实时监控举升任务的简化流程,在视频中识别举升动作并计算举升次数。背景:识别举重动作的传统方法通常依赖于深度学习分类器,该分类器应用于从可穿戴传感器收集的人体运动数据。尽管这些方法具有高性能,但在硬件资源有限的系统上很难实现。方法:该方法分为五个阶段:(1)实时姿态估计模型BlazePose检测人体关键关节。(2)通过平滑、定心和缩放技术对这些接头进行预处理。(3)提取预处理后关节的运动特征。(4)利用秩变运动特征对对视频帧进行升降和非升降分类。(5)提升计数算法根据类别预测对提升次数进行计数。结果:确定了9个秩变的运动特征对作为关键对。使用这些对构建一个集成分类器,该分类器在分类指标上达到0.89或以上,包括准确率、精密度、召回率和F1分数。该分类器的提升计数准确率为0.90,延迟为0.06 ms,比基线分类器至少快12.5倍。结论:本研究表明,基于计算机视觉的运动学特征可以有效地识别举重动作。应用:所提出的方法可以部署在各种平台上,包括移动设备和嵌入式系统,以实时监控举升任务,主动预防与工作相关的腰背部伤害。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Human Factors
Human Factors 管理科学-行为科学
CiteScore
10.60
自引率
6.10%
发文量
99
审稿时长
6-12 weeks
期刊介绍: Human Factors: The Journal of the Human Factors and Ergonomics Society publishes peer-reviewed scientific studies in human factors/ergonomics that present theoretical and practical advances concerning the relationship between people and technologies, tools, environments, and systems. Papers published in Human Factors leverage fundamental knowledge of human capabilities and limitations – and the basic understanding of cognitive, physical, behavioral, physiological, social, developmental, affective, and motivational aspects of human performance – to yield design principles; enhance training, selection, and communication; and ultimately improve human-system interfaces and sociotechnical systems that lead to safer and more effective outcomes.
期刊最新文献
Human-Robot Collaboration With a Corrective Shared Controlled Robot in a Sanding Task. Gaze Sharing, a Double-Edged Sword: Examining the Effect of Real-Time Gaze Sharing Visualizations on Team Performance and Situation Awareness. Hello, is it me you're Stopping for? The Effect of external Human Machine Interface Familiarity on Pedestrians' Crossing Behaviour in an Ambiguous Situation. Multitasking Induced Contextual Blindness. Deciphering Automation Transparency: Do the Benefits of Transparency Differ Based on Whether Decision Recommendations Are Provided?
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1