As an intuitive form of interaction, gestures are widely used in fields like smart live streaming and smart homes, significantly improving user efficiency and the convenience of human-computer interaction. However, existing gesture recognition methods often suffer from large model sizes and high computational complexity, making them unsuitable for real-time deployment. To address these issues, we propose GF-YOLOv8, a lightweight gesture recognition model based on YOLOv8 (where ‘GF’ stands for the initials of two optimization modules: G represents the GB-C2f module, and F represents the FFPN network), designed to reduce parameters while maintaining detection accuracy. First, by designing the GB-C2f module, which integrates the lightweight network GhostNet, the model significantly reduces computation and parameters with minimal loss in accuracy. Second, a fusion feature pyramid network (FFPN), based on ASFF and AFPN algorithms, and enhances the interaction between non-adjacent layers to improve gesture feature perception. Compared to the original YOLOv8 model, the proposed method demonstrates significant improvements on both public and self-constructed datasets: GFLOPs and parameter count are reduced by 46.9% and 44.9%, respectively, with the mAP value remaining almost unchanged. The improved model enables efficient and accurate gesture recognition on mobile and embedded devices, achieving a substantial reduction in model size without compromising accuracy. © 2025 Institute of Electrical Engineers of Japan and Wiley Periodicals LLC.
{"title":"GF-YOLOv8: A Lightweight Gesture Recognition Algorithm Based on Adaptive Feature Fusion Pyramid Network","authors":"Yiqing Liu, Linxiao Zheng, Miao Wu, Lin Wang, Lin Zhou","doi":"10.1002/tee.70134","DOIUrl":"https://doi.org/10.1002/tee.70134","url":null,"abstract":"<p>As an intuitive form of interaction, gestures are widely used in fields like smart live streaming and smart homes, significantly improving user efficiency and the convenience of human-computer interaction. However, existing gesture recognition methods often suffer from large model sizes and high computational complexity, making them unsuitable for real-time deployment. To address these issues, we propose GF-YOLOv8, a lightweight gesture recognition model based on YOLOv8 (where ‘GF’ stands for the initials of two optimization modules: G represents the GB-C2f module, and F represents the FFPN network), designed to reduce parameters while maintaining detection accuracy. First, by designing the GB-C2f module, which integrates the lightweight network GhostNet, the model significantly reduces computation and parameters with minimal loss in accuracy. Second, a fusion feature pyramid network (FFPN), based on ASFF and AFPN algorithms, and enhances the interaction between non-adjacent layers to improve gesture feature perception. Compared to the original YOLOv8 model, the proposed method demonstrates significant improvements on both public and self-constructed datasets: GFLOPs and parameter count are reduced by 46.9% and 44.9%, respectively, with the mAP value remaining almost unchanged. The improved model enables efficient and accurate gesture recognition on mobile and embedded devices, achieving a substantial reduction in model size without compromising accuracy. © 2025 Institute of Electrical Engineers of Japan and Wiley Periodicals LLC.</p>","PeriodicalId":13435,"journal":{"name":"IEEJ Transactions on Electrical and Electronic Engineering","volume":"21 1","pages":"106-113"},"PeriodicalIF":1.1,"publicationDate":"2025-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145754569","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0