The Use of Video Captioning for Fostering Physical Activity

Soheyla Amirian, Abolfazl Farahani, H. Arabnia, K. Rasheed, T. Taha
{"title":"The Use of Video Captioning for Fostering Physical Activity","authors":"Soheyla Amirian, Abolfazl Farahani, H. Arabnia, K. Rasheed, T. Taha","doi":"10.1109/CSCI51800.2020.00108","DOIUrl":null,"url":null,"abstract":"Video Captioning is considered to be one of the most challenging problems in the field of computer vision. Video Captioning involves the combination of different deep learning models to perform object detection, action detection, and localization by processing a sequence of image frames. It is crucial to consider the sequence of actions in a video in order to generate a meaningful description of the overall action event. A reliable, accurate, and real-time video captioning method can be used in many applications. However, this paper focuses on one application: video captioning for fostering and facilitating physical activities. In broad terms, the work can be considered to be assistive technology. Lack of physical activity appears to be increasingly widespread in many nations due to many factors, the most important being the convenience that technology has provided in workplaces. The adopted sedentary lifestyle is becoming a significant public health issue. Therefore, it is essential to incorporate more physical movements into our daily lives. Tracking one’s daily physical activities would offer a base for comparison with activities performed in subsequent days. With the above in mind, this paper proposes a video captioning framework that aims to describe the activities in a video and estimate a person’s daily physical activity level. This framework could potentially help people trace their daily movements to reduce an inactive lifestyle’s health risks. The work presented in this paper is still in its infancy. The initial steps of the application are outlined in this paper. Based on our preliminary research, this project has great merit.","PeriodicalId":336929,"journal":{"name":"2020 International Conference on Computational Science and Computational Intelligence (CSCI)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 International Conference on Computational Science and Computational Intelligence (CSCI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CSCI51800.2020.00108","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 8

Abstract

Video Captioning is considered to be one of the most challenging problems in the field of computer vision. Video Captioning involves the combination of different deep learning models to perform object detection, action detection, and localization by processing a sequence of image frames. It is crucial to consider the sequence of actions in a video in order to generate a meaningful description of the overall action event. A reliable, accurate, and real-time video captioning method can be used in many applications. However, this paper focuses on one application: video captioning for fostering and facilitating physical activities. In broad terms, the work can be considered to be assistive technology. Lack of physical activity appears to be increasingly widespread in many nations due to many factors, the most important being the convenience that technology has provided in workplaces. The adopted sedentary lifestyle is becoming a significant public health issue. Therefore, it is essential to incorporate more physical movements into our daily lives. Tracking one’s daily physical activities would offer a base for comparison with activities performed in subsequent days. With the above in mind, this paper proposes a video captioning framework that aims to describe the activities in a video and estimate a person’s daily physical activity level. This framework could potentially help people trace their daily movements to reduce an inactive lifestyle’s health risks. The work presented in this paper is still in its infancy. The initial steps of the application are outlined in this paper. Based on our preliminary research, this project has great merit.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
使用视频字幕促进体育活动
视频字幕被认为是计算机视觉领域最具挑战性的问题之一。视频字幕涉及不同深度学习模型的组合,通过处理一系列图像帧来执行对象检测、动作检测和定位。考虑视频中的行动顺序是至关重要的,这样才能对整个行动事件产生有意义的描述。一种可靠、准确、实时的视频字幕方法可用于多种应用。然而,本文的重点是一个应用:视频字幕促进和促进体育活动。从广义上讲,这项工作可以被认为是辅助技术。由于许多因素,缺乏体育锻炼似乎在许多国家越来越普遍,最重要的是技术在工作场所提供的便利。久坐不动的生活方式正在成为一个重大的公共健康问题。因此,在我们的日常生活中加入更多的体育运动是必不可少的。跟踪一个人的日常体育活动将为与随后几天的活动进行比较提供一个基础。基于上述考虑,本文提出了一个视频字幕框架,旨在描述视频中的活动并估计一个人的日常身体活动水平。这个框架可能会帮助人们追踪他们的日常活动,以减少不活跃的生活方式带来的健康风险。本文所介绍的工作仍处于起步阶段。本文概述了应用程序的初始步骤。根据我们的初步研究,这个项目有很大的优点。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
First Success of Cancer Gene Data Analysis of 169 Microarrays for Medical Diagnosis Artificial Intelligence in Computerized Adaptive Testing Evidence for Monitoring the User and Computing the User’s trust Transfer of Hierarchical Reinforcement Learning Structures for Robotic Manipulation Tasks An open-source application built with R programming language for clinical laboratories to innovate process of excellence and overcome the uncertain outlook during the global healthcare crisis
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1