首页 > 最新文献

Computers and Electronics in Agriculture最新文献

英文 中文
GLS-YOLOv8n: a lightweight ‘Guiqi’ mango detection model via RGB-depth-thermal image fusion GLS-YOLOv8n:一种基于rgb -深度-热图像融合的轻型“桂气”芒果检测模型
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-20 DOI: 10.1016/j.compag.2025.111355
Weihua Cao , Zhao Zhang , Zeping Wang , Ailin Wei , Qianfu Chen , C. Igathinathane , Fu Zhang , Mahmoud A. Abdelhamid , Dapeng Ye , Yiannis Ampatzidis
Accurate and efficient fruit detection and localization are essential for the development of automated harvesting systems. Existing mango detection approaches encounter challenges in complex orchard conditions, including variable lighting, foliage occlusion, small targets, and fruit overlapping. To address these challenges, this study developed a robust and lightweight detection model that maintains both high accuracy and computational efficiency, making it suitable for real-time applications. To support model training and validation, the dataset was collected from a mango orchard under diverse illumination and occlusion scenarios, comprising 353 sets of synchronized RGB, depth, and thermal images. To leverage these multimodal data, a fusion strategy was proposed by integrating RGB (textural feature), depth (spatial structure feature), and thermal images (temperature feature) to exploit their complementary strengths. Experimental results using YOLOv8n as the baseline demonstrated that trimodal fusion significantly outperformed single-modality inputs, achieving a 97.2 % average precision (AP), which was 2.4 % higher than the best single-modality. Based on this, GLS-YOLOv8n was proposed by incorporating GhostHGNetv2 as a lightweight backbone, a lightweight shared convolutional detection head (Detect-LSCD) for efficient small-object detection, and C2f-Star module for optimized multimodal feature fusion. At a speed of 65.7 fps, GLS-YOLOv8n achieved an AP of 98.5 % by reducing the parameter size from 3.0 M to 1.4 M (53 % reduction), floating-point operations (FLOPs) from 8.2 G to 5.0 G (39 % reduction), and compressing the model size from 5.98 to 3.06 MB (49 % reduction). The findings of this study demonstrated that GLS-YOLOv8n achieved a good balance between accuracy and efficiency, making it suitable for real-time mango detection under natural environments.
准确、高效的水果检测和定位是自动化采收系统发展的必要条件。现有的芒果检测方法在复杂的果园条件下遇到挑战,包括可变光照、树叶遮挡、小目标和水果重叠。为了应对这些挑战,本研究开发了一种鲁棒且轻量级的检测模型,该模型保持了高精度和计算效率,使其适合于实时应用。为了支持模型训练和验证,数据集采集于不同光照和遮挡场景下的芒果果园,包括353组同步的RGB、深度和热图像。为了利用这些多模态数据,提出了一种融合RGB(纹理特征)、深度(空间结构特征)和热图像(温度特征)的融合策略,以利用它们的互补优势。以YOLOv8n为基准的实验结果表明,三模态融合显著优于单模态输入,达到97.2%的平均精度(AP),比最佳单模态高2.4%。在此基础上,结合GhostHGNetv2作为轻量级骨干,利用轻量级共享卷积检测头(Detect-LSCD)高效检测小目标,利用C2f-Star模块优化多模态特征融合,提出了GLS-YOLOv8n算法。在65.7 fps的速度下,GLS-YOLOv8n通过将参数大小从3.0 M减少到1.4 M(减少53%),浮点运算(FLOPs)从8.2 G减少到5.0 G(减少39%),将模型大小从5.98压缩到3.06 MB(减少49%),实现了98.5%的AP。本研究结果表明,GLS-YOLOv8n在准确性和效率之间取得了很好的平衡,适用于自然环境下的芒果实时检测。
{"title":"GLS-YOLOv8n: a lightweight ‘Guiqi’ mango detection model via RGB-depth-thermal image fusion","authors":"Weihua Cao ,&nbsp;Zhao Zhang ,&nbsp;Zeping Wang ,&nbsp;Ailin Wei ,&nbsp;Qianfu Chen ,&nbsp;C. Igathinathane ,&nbsp;Fu Zhang ,&nbsp;Mahmoud A. Abdelhamid ,&nbsp;Dapeng Ye ,&nbsp;Yiannis Ampatzidis","doi":"10.1016/j.compag.2025.111355","DOIUrl":"10.1016/j.compag.2025.111355","url":null,"abstract":"<div><div>Accurate and efficient fruit detection and localization are essential for the development of automated harvesting systems. Existing mango detection approaches encounter challenges in complex orchard conditions, including variable lighting, foliage occlusion, small targets, and fruit overlapping. To address these challenges, this study developed a robust and lightweight detection model that maintains both high accuracy and computational efficiency, making it suitable for real-time applications. To support model training and validation, the dataset was collected from a mango orchard under diverse illumination and occlusion scenarios, comprising 353 sets of synchronized RGB, depth, and thermal images. To leverage these multimodal data, a fusion strategy was proposed by integrating RGB (textural feature), depth (spatial structure feature), and thermal images (temperature feature) to exploit their complementary strengths. Experimental results using YOLOv8n as the baseline demonstrated that trimodal fusion significantly outperformed single-modality inputs, achieving a 97.2 % average precision (AP), which was 2.4 % higher than the best single-modality. Based on this, GLS-YOLOv8n was proposed by incorporating GhostHGNetv2 as a lightweight backbone, a lightweight shared convolutional detection head (Detect-LSCD) for efficient small-object detection, and C2f-Star module for optimized multimodal feature fusion. At a speed of 65.7 fps, GLS-YOLOv8n achieved an AP of 98.5 % by reducing the parameter size from 3.0 M to 1.4 M (53 % reduction), floating-point operations (FLOPs) from 8.2 G to 5.0 G (39 % reduction), and compressing the model size from 5.98 to 3.06 MB (49 % reduction). The findings of this study demonstrated that GLS-YOLOv8n achieved a good balance between accuracy and efficiency, making it suitable for real-time mango detection under natural environments.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"242 ","pages":"Article 111355"},"PeriodicalIF":8.9,"publicationDate":"2025-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145791866","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dynamic light intensity improves light use efficiency of lettuce in vertical farming: quantifying light interception through 3D phenotyping analysis 动态光强提高垂直种植莴苣的光利用效率:通过三维表型分析量化光截获
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-20 DOI: 10.1016/j.compag.2025.111349
Zhixian Lin, Xunyi Ma, Wei Liu, Tao Lin
Vertical farming offers a promising solution to global food security and urbanization challenges, yet its widespread adoption is hindered by high costs, particularly for lighting. Addressing this requires enhancing light use efficiency (LUE) through intelligent control strategies. While numerous studies have investigated the effects of light intensity on lettuce growth, relatively few have explored the potential benefits of stage-specific light regulation. In this study, we first developed an automated 3D phenotyping pipeline based on multi-view reconstruction to quantify canopy morphology and light interception. Utilizing this quantitative framework, we conducted a dynamic light experiment with lettuce in a commercial plant factory to evaluate four dynamic light-intensity strategies. The proposed 3D phenotyping pipeline demonstrated promising performance for canopy information extraction, with RMSEs for plant height, canopy diameter, and projected leaf area of 0.79 cm, 1.05 cm, and 44.3 cm2, respectively. The “high-low-high” dynamic lighting strategy, applying higher light intensity during the early and late growth stages and lower intensity during the mid-growth stage, successfully optimized canopy morphology for better light capture. This treatment significantly increased shoot fresh and dry weights by 28 % and 65 %, respectively, compared to constant lighting. Furthermore, it enhanced LUE based on incident and intercepted light integrals by 67 % and 19 %, while reducing electricity consumption per unit of fresh weight by 24 %. Nutritional quality analysis showed the treatment increased soluble sugars and starch contents. By integrating advanced 3D phenotyping with dynamic light intensity control, this study demonstrates a prototype for intelligent decision-making to enhance yield and energy use efficiency in practical vertical farming.
垂直农业为全球粮食安全和城市化挑战提供了一个有希望的解决方案,但其广泛采用受到高成本的阻碍,特别是照明成本高。解决这个问题需要通过智能控制策略提高光的使用效率(LUE)。虽然有许多研究调查了光照强度对生菜生长的影响,但相对较少的研究探索了特定阶段光照调节的潜在益处。在这项研究中,我们首先开发了一个基于多视图重建的自动化3D表型管道,以量化冠层形态和光拦截。利用这一定量框架,我们在一家商业植物工厂对生菜进行了动态光实验,以评估四种动态光强策略。所提出的三维表型管道在提取冠层信息方面表现出良好的性能,植物高度、冠层直径和投影叶面积的rmse分别为0.79 cm、1.05 cm和44.3 cm2。采用“高-低-高”的动态光照策略,在生长前期和后期施加较高的光照强度,在生长中期施加较低的光照强度,成功地优化了冠层形态,实现了更好的光捕获。与恒定光照相比,该处理显著提高了地上部鲜重和干重,分别提高了28%和65%。此外,它还将入射光积分和拦截光积分的LUE分别提高了67%和19%,同时将单位鲜重的电力消耗降低了24%。营养品质分析表明,处理提高了可溶性糖和淀粉含量。通过将先进的3D表型与动态光强控制相结合,本研究展示了一个智能决策的原型,以提高实际垂直农业的产量和能源利用效率。
{"title":"Dynamic light intensity improves light use efficiency of lettuce in vertical farming: quantifying light interception through 3D phenotyping analysis","authors":"Zhixian Lin,&nbsp;Xunyi Ma,&nbsp;Wei Liu,&nbsp;Tao Lin","doi":"10.1016/j.compag.2025.111349","DOIUrl":"10.1016/j.compag.2025.111349","url":null,"abstract":"<div><div>Vertical farming offers a promising solution to global food security and urbanization challenges, yet its widespread adoption is hindered by high costs, particularly for lighting. Addressing this requires enhancing light use efficiency (LUE) through intelligent control strategies. While numerous studies have investigated the effects of light intensity on lettuce growth, relatively few have explored the potential benefits of stage-specific light regulation. In this study, we first developed an automated 3D phenotyping pipeline based on multi-view reconstruction to quantify canopy morphology and light interception. Utilizing this quantitative framework, we conducted a dynamic light experiment with lettuce in a commercial plant factory to evaluate four dynamic light-intensity strategies. The proposed 3D phenotyping pipeline demonstrated promising performance for canopy information extraction, with RMSEs for plant height, canopy diameter, and projected leaf area of 0.79 cm, 1.05 cm, and 44.3 cm<sup>2</sup>, respectively. The “high-low-high” dynamic lighting strategy, applying higher light intensity during the early and late growth stages and lower intensity during the mid-growth stage, successfully optimized canopy morphology for better light capture. This treatment significantly increased shoot fresh and dry weights by 28 % and 65 %, respectively, compared to constant lighting. Furthermore, it enhanced LUE based on incident and intercepted light integrals by 67 % and 19 %, while reducing electricity consumption per unit of fresh weight by 24 %. Nutritional quality analysis showed the treatment increased soluble sugars and starch contents. By integrating advanced 3D phenotyping with dynamic light intensity control, this study demonstrates a prototype for intelligent decision-making to enhance yield and energy use efficiency in practical vertical farming.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"242 ","pages":"Article 111349"},"PeriodicalIF":8.9,"publicationDate":"2025-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145791867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A magnetic induction network for high-resolution, real-time soil moisture monitoring in complex subsurface environments 用于复杂地下环境中高分辨率、实时土壤湿度监测的磁感应网络
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-20 DOI: 10.1016/j.compag.2025.111314
Zhangyu Li
Water resource management in smart agriculture demands accurate, real-time soil moisture monitoring across large and heterogeneous fields. However, existing sensing technologies, such as time-domain reflectometry (TDR), capacitance probes, ground-penetrating radar (GPR), and single-link wireless underground systems, suffer from limited spatial coverage, reduced performance in conductive soils, and an inability to resolve heterogeneous moisture patterns in real time. This study aims to develop and validate a scalable magnetic induction (MI) network–based sensing system that enables real-time, high-resolution volumetric soil moisture mapping across large and heterogeneous agricultural fields, thereby overcoming the limited coverage and instability of conventional underground sensing technologies. In particular, the system combines three main elements: (i) a distributed magnetic induction (MI) networking architecture designed for reliable underground communication; (ii) a channel frequency response (CFR)-based protocol that supports real-time multi-link operations; and (iii) a multi-frequency phase unwrapping algorithm that ensures precise phase reconstruction across diverse soil conditions. Comparative evaluation through COMSOL Multiphysics full-wave simulations shows that the system reduces average sensing errors by 25–50 % over a 40 m range compared with conventional wireless underground sensing. Laboratory testbed experiments confirm a 38 % accuracy improvement over state-of-the-art approaches, achieving a spatial RMSE of ∼0.02 m3/m3. This work is the first experimental demonstration of using MI networking for 3D real-time soil moisture imaging, offering a robust and scalable solution for precision agriculture. By enabling distributed, non-invasive subsurface monitoring, the system has the potential to improve irrigation scheduling and enhance water-use efficiency by an estimated 15–25 % in large-scale farming operations.
智能农业中的水资源管理需要对大型异质农田进行精确、实时的土壤湿度监测。然而,现有的传感技术,如时域反射仪(TDR)、电容探头、探地雷达(GPR)和单链路无线地下系统,存在空间覆盖有限、在导电土壤中的性能下降以及无法实时解决非均匀水分模式的问题。本研究旨在开发和验证一种可扩展的基于磁感应(MI)网络的传感系统,该系统能够在大型异质农业领域实现实时、高分辨率的体积土壤湿度测绘,从而克服传统地下传感技术的有限覆盖范围和不稳定性。特别是,该系统结合了三个主要元素:(i)为可靠的地下通信设计的分布式磁感应(MI)网络架构;(ii)支持实时多链路操作的基于信道频率响应(CFR)的协议;(iii)确保在不同土壤条件下精确重建相位的多频率相位展开算法。通过COMSOL多物理场全波模拟的对比评估表明,与传统的无线地下传感相比,该系统在40 m范围内的平均传感误差降低了25-50 %。实验室试验台实验证实,与最先进的方法相比,精确度提高了38 %,实现了 ~ 0.02 m3/m3的空间RMSE。这项工作是使用MI网络进行三维实时土壤水分成像的第一次实验演示,为精准农业提供了强大且可扩展的解决方案。通过实现分布式、非侵入式地下监测,该系统有可能改善灌溉计划,并在大规模农业经营中提高用水效率,估计可提高15 - 25% %。
{"title":"A magnetic induction network for high-resolution, real-time soil moisture monitoring in complex subsurface environments","authors":"Zhangyu Li","doi":"10.1016/j.compag.2025.111314","DOIUrl":"10.1016/j.compag.2025.111314","url":null,"abstract":"<div><div>Water resource management in smart agriculture demands accurate, real-time soil moisture monitoring across large and heterogeneous fields. However, existing sensing technologies, such as time-domain reflectometry (TDR), capacitance probes, ground-penetrating radar (GPR), and single-link wireless underground systems, suffer from limited spatial coverage, reduced performance in conductive soils, and an inability to resolve heterogeneous moisture patterns in real time. This study aims to develop and validate a scalable magnetic induction (MI) network–based sensing system that enables real-time, high-resolution volumetric soil moisture mapping across large and heterogeneous agricultural fields, thereby overcoming the limited coverage and instability of conventional underground sensing technologies. In particular, the system combines three main elements: (i) a distributed magnetic induction (MI) networking architecture designed for reliable underground communication; (ii) a channel frequency response (CFR)-based protocol that supports real-time multi-link operations; and (iii) a multi-frequency phase unwrapping algorithm that ensures precise phase reconstruction across diverse soil conditions. Comparative evaluation through COMSOL Multiphysics full-wave simulations shows that the system reduces average sensing errors by 25–50 % over a 40 m range compared with conventional wireless underground sensing. Laboratory testbed experiments confirm a 38 % accuracy improvement over state-of-the-art approaches, achieving a spatial RMSE of ∼0.02 m<sup>3</sup>/m<sup>3</sup>. This work is the first experimental demonstration of using MI networking for 3D real-time soil moisture imaging, offering a robust and scalable solution for precision agriculture. By enabling distributed, non-invasive subsurface monitoring, the system has the potential to improve irrigation scheduling and enhance water-use efficiency by an estimated 15–25 % in large-scale farming operations.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"242 ","pages":"Article 111314"},"PeriodicalIF":8.9,"publicationDate":"2025-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145791865","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Scalable phenotyping and yield estimation via stability index and single-plant variability using a vision-based large model framework 使用基于视觉的大型模型框架,通过稳定性指数和单株变异进行可扩展表型和产量估计
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-19 DOI: 10.1016/j.compag.2025.111300
Mian Chen , Guangyao Sun , Kai Ma , Xiaoli Geng , Daowu Hu , ShuaiPeng Fei , Xiongming Du , Shoupu He , Rui Zhang , Shunfu Xiao , Lei Meng , Yuntao Ma
High-throughput field phenotyping bridges genotype, environment, and phenotypic performance. Conventional plot-level approaches relying on manual surveys are labor-intensive, and error-prone and fail to capture variability among individual plants, limiting seed cotton yield estimation and genotype screening under natural conditions. To address these limitations, a complex framework was developed, integrating single-plant instance segmentation, multi-trait inversion, plot-level stability characterization, and yield estimation. The enhanced vision-model framework, TopoRefineSAM, combines YOLOv12 detection with SAM2 segmentation and incorporates adaptive enhancement and topological refinement modules, enabling efficient, robust, and cost-effective single-plant identification under weak annotation. Based on this segmentation, multi-source UAV imagery (RGB, multispectral, thermal infrared, and DSM) was used to build ensemble learning models for inversion of physiological and biomass traits. A Stability Index Group (SIG) translates inter-plant variability into plot-level stability features, improving interpretability and consistency in yield estimation and cultivar screening. Results demonstrated that TopoRefineSAM achieved high segmentation accuracy for single-plant extraction under complex field conditions. In multi-trait inversion, Gradient Boosting Decision Trees (GBDT) achieved the highest performance. Our results demonstrated strong consistency between multimodal features and measured traits. In yield estimation, incorporating the SIG substantially improved predictive performance across growth stages. In cultivar screening, the method achieved high agreement with field measurements, showing robust identification of top-performing cultivars. Collectively, the findings establish a scalable, cost-effective, and high-accuracy framework for field-based phenotypic analysis and yield estimation, providing both methodological innovations and practical support for precision breeding and large-scale crop improvement.
高通量现场表型连接基因型,环境和表型性能。传统的依赖人工调查的地块水平方法是劳动密集型的,容易出错,并且不能捕获单个植物之间的变异性,限制了自然条件下棉籽产量估计和基因型筛选。为了解决这些限制,开发了一个复杂的框架,集成了单株实例分割,多性状反演,地块稳定性表征和产量估计。增强的视觉模型框架TopoRefineSAM结合了YOLOv12检测和SAM2分割,并结合了自适应增强和拓扑细化模块,实现了弱注释下高效、鲁棒和经济的单植物识别。在此基础上,利用多源无人机图像(RGB、多光谱、热红外和DSM)构建生理和生物量特征反演的集成学习模型。稳定性指数组(SIG)将植物间变异转化为地块水平的稳定性特征,提高产量估算和品种筛选的可解释性和一致性。结果表明,在复杂的野外条件下,TopoRefineSAM对单一植物的提取具有较高的分割精度。在多特征反演中,梯度增强决策树(GBDT)获得了最高的性能。我们的结果表明,多模态特征和测量特征之间存在很强的一致性。在产量估计中,结合SIG大大提高了整个生长阶段的预测性能。在品种筛选中,该方法与田间测量结果的一致性较高,对优良品种的识别效果较好。总的来说,这些发现为基于田间的表型分析和产量估计建立了一个可扩展的、具有成本效益的和高精度的框架,为精确育种和大规模作物改良提供了方法创新和实践支持。
{"title":"Scalable phenotyping and yield estimation via stability index and single-plant variability using a vision-based large model framework","authors":"Mian Chen ,&nbsp;Guangyao Sun ,&nbsp;Kai Ma ,&nbsp;Xiaoli Geng ,&nbsp;Daowu Hu ,&nbsp;ShuaiPeng Fei ,&nbsp;Xiongming Du ,&nbsp;Shoupu He ,&nbsp;Rui Zhang ,&nbsp;Shunfu Xiao ,&nbsp;Lei Meng ,&nbsp;Yuntao Ma","doi":"10.1016/j.compag.2025.111300","DOIUrl":"10.1016/j.compag.2025.111300","url":null,"abstract":"<div><div>High-throughput field phenotyping bridges genotype, environment, and phenotypic performance. Conventional plot-level approaches relying on manual surveys are labor-intensive, and error-prone and fail to capture variability among individual plants, limiting seed cotton yield estimation and genotype screening under natural conditions. To address these limitations, a complex framework was developed, integrating single-plant instance segmentation, multi-trait inversion, plot-level stability characterization, and yield estimation. The enhanced vision-model framework, TopoRefineSAM, combines YOLOv12 detection with SAM2 segmentation and incorporates adaptive enhancement and topological refinement modules, enabling efficient, robust, and cost-effective single-plant identification under weak annotation. Based on this segmentation, multi-source UAV imagery (RGB, multispectral, thermal infrared, and DSM) was used to build ensemble learning models for inversion of physiological and biomass traits. A Stability Index Group (SIG) translates inter-plant variability into plot-level stability features, improving interpretability and consistency in yield estimation and cultivar screening. Results demonstrated that TopoRefineSAM achieved high segmentation accuracy for single-plant extraction under complex field conditions. In multi-trait inversion, Gradient Boosting Decision Trees (GBDT) achieved the highest performance. Our results demonstrated strong consistency between multimodal features and measured traits. In yield estimation, incorporating the SIG substantially improved predictive performance across growth stages. In cultivar screening, the method achieved high agreement with field measurements, showing robust identification of top-performing cultivars. Collectively, the findings establish a scalable, cost-effective, and high-accuracy framework for field-based phenotypic analysis and yield estimation, providing both methodological innovations and practical support for precision breeding and large-scale crop improvement.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"242 ","pages":"Article 111300"},"PeriodicalIF":8.9,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145791869","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Grower-in-the-loop interactive reinforcement learning for greenhouse climate control 用于温室气候控制的生长者在环交互式强化学习
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-18 DOI: 10.1016/j.compag.2025.111312
Maxiu Xiao , Jianglin Lan , Jingxin Yu , Weihong Ma , Qiuju Xie , Congcong Sun
Climate control is crucial for greenhouse production as it directly affects crop growth and resource use. Reinforcement learning (RL) has received increasing attention in this field, but still faces challenges, including limited training efficiency and high reliance on initial learning conditions. Interactive RL, which integrates human (grower) input with the RL agent’s learning, offers a potential solution to overcome these challenges. However, interactive RL has not yet been applied to greenhouse climate control. Furthermore, human input is hardly perfect due to the complexity of climate control. The performance of interactive RL may also be limited by imperfect input. Therefore, this paper aims to explore the possibility and performance of applying interactive RL with imperfect inputs into greenhouse climate control, by: (1) developing three representative interactive RL algorithms tailored for greenhouse climate control (reward shaping, policy shaping and control sharing); (2) analyzing how input characteristics are often contradicting, and how the trade-offs between them make grower’s inputs difficult to perfect; (3) proposing a neural network-based approach to enhance the robustness of interactive RL agents under limited input availability; (4) conducting a comprehensive evaluation of the three interactive RL algorithms with imperfect inputs in a simulated greenhouse environment. The demonstration shows that interactive RL incorporating imperfect grower inputs has the potential to improve the performance of the RL agent. Interactive RL algorithms that influence action selection, such as policy shaping and control sharing, perform better when dealing with imperfect inputs, achieving 8.4% and 6.8% improvement in profit, respectively. In contrast, reward shaping, an algorithm that manipulates the reward function, is sensitive to imperfect inputs and leads to a 9.4% decrease in profit. This highlights the importance of selecting an appropriate mechanism when incorporating imperfect inputs.
气候控制对温室生产至关重要,因为它直接影响作物生长和资源利用。强化学习(RL)在该领域受到越来越多的关注,但仍然面临着训练效率有限和对初始学习条件的高度依赖等挑战。交互式强化学习将人类(种植者)的输入与强化学习代理的学习相结合,为克服这些挑战提供了一个潜在的解决方案。然而,交互式RL尚未应用于温室气候控制。此外,由于气候控制的复杂性,人类的投入并不完美。交互式强化学习的性能也可能受到不完善输入的限制。因此,本文旨在探讨将不完全输入交互式强化学习应用于温室气候控制的可能性和性能,方法如下:(1)针对温室气候控制开发三种具有代表性的交互式强化学习算法(奖励塑造、政策塑造和控制权共享);(2)分析投入品特征如何经常相互矛盾,以及它们之间的权衡如何使种植者的投入品难以完善;(3)提出了一种基于神经网络的增强交互式RL智能体在有限输入可用性下的鲁棒性的方法;(4)在模拟温室环境下,对三种输入不完全的交互式强化学习算法进行了综合评价。该演示表明,包含不完美种植者输入的交互式强化学习有可能提高强化学习代理的性能。影响行为选择的交互式强化学习算法,如策略形成和控制共享,在处理不完美输入时表现更好,分别实现了8.4%和6.8%的利润提高。相比之下,奖励塑造(reward shaping),一种操纵奖励函数的算法,对不完美的输入很敏感,导致利润下降9.4%。这突出了在纳入不完善的投入时选择适当机制的重要性。
{"title":"Grower-in-the-loop interactive reinforcement learning for greenhouse climate control","authors":"Maxiu Xiao ,&nbsp;Jianglin Lan ,&nbsp;Jingxin Yu ,&nbsp;Weihong Ma ,&nbsp;Qiuju Xie ,&nbsp;Congcong Sun","doi":"10.1016/j.compag.2025.111312","DOIUrl":"10.1016/j.compag.2025.111312","url":null,"abstract":"<div><div>Climate control is crucial for greenhouse production as it directly affects crop growth and resource use. Reinforcement learning (RL) has received increasing attention in this field, but still faces challenges, including limited training efficiency and high reliance on initial learning conditions. Interactive RL, which integrates human (grower) input with the RL agent’s learning, offers a potential solution to overcome these challenges. However, interactive RL has not yet been applied to greenhouse climate control. Furthermore, human input is hardly perfect due to the complexity of climate control. The performance of interactive RL may also be limited by imperfect input. Therefore, this paper aims to explore the possibility and performance of applying interactive RL with imperfect inputs into greenhouse climate control, by: (1) developing three representative interactive RL algorithms tailored for greenhouse climate control (reward shaping, policy shaping and control sharing); (2) analyzing how input characteristics are often contradicting, and how the trade-offs between them make grower’s inputs difficult to perfect; (3) proposing a neural network-based approach to enhance the robustness of interactive RL agents under limited input availability; (4) conducting a comprehensive evaluation of the three interactive RL algorithms with imperfect inputs in a simulated greenhouse environment. The demonstration shows that interactive RL incorporating imperfect grower inputs has the potential to improve the performance of the RL agent. Interactive RL algorithms that influence action selection, such as policy shaping and control sharing, perform better when dealing with imperfect inputs, achieving 8.4% and 6.8% improvement in profit, respectively. In contrast, reward shaping, an algorithm that manipulates the reward function, is sensitive to imperfect inputs and leads to a 9.4% decrease in profit. This highlights the importance of selecting an appropriate mechanism when incorporating imperfect inputs.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"242 ","pages":"Article 111312"},"PeriodicalIF":8.9,"publicationDate":"2025-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145791789","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research on real time mapping and transfer learning based 3D semantic segmentation of unstructured orchards 基于实时映射和迁移学习的非结构化果园三维语义分割研究
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-18 DOI: 10.1016/j.compag.2025.111323
Sipei Li , Sheng Wen , Fagang Liu , Haiyan Zhang , Yufeng Ge , Zhiyin Deng
Precise semantic maps can provide considerable assistance in the intelligent management of unstructured orchards. However, successful extraction of useful tree information from complex environments to achieve autonomous unmanned aerial vehicle (UAV) operations in modern smart agriculture yet to be achieved.
Herein, we propose a real-time mapping and transfer learning-based 3D semantic segmentation framework to enable autonomous UAV operations in orchards. First, we constructed an autonomous perception platform. This platform integrates an optimized light detection and ranging–simultaneous localization and mapping framework. The framework fuses point cloud data from light detection and ranging (LiDAR), pose data from inertial measurement units (IMU), and Global Navigation satellite system (GNSS) signals. As a result, it can accurately capture point clouds of objects in unstructured orchards. Second, using transfer learning, we achieved rapid semantic segmentation of orchard point cloud maps with a small sample dataset. Finally, we employed a clustering segmentation algorithm to isolate individual lemon trees and extract parameters such as tree height, tree center location, and canopy size. The method proposed in this study was evaluated in a real-world scenario of an unstructured lemon orchard. The proposed LiDAR SLAM framework demonstrated a significant reduction in vertical root mean square error (RMSE). The framework achieved RMSE reductions of 68.55 % and 53.15 % against FAST-LIO2 and LIO-SAM, respectively. The transfer learning-based segmentation achieved a mean intersection over union (mIoU) 3.6 % higher than the baseline PointNet++. Furthermore, the extracted tree parameters exhibited a strong correlation with ground-truth measurements. This work provides a validated end-to-end solution for automated orchard perception, offering key data for precision management tasks.
精确的语义地图可以为非结构化果园的智能管理提供相当大的帮助。然而,在现代智慧农业中,从复杂环境中成功提取有用的树木信息以实现无人驾驶飞行器(UAV)的自主操作尚未实现。在此,我们提出了一个基于实时映射和迁移学习的3D语义分割框架,以实现无人机在果园中的自主操作。首先,我们构建了一个自主感知平台。该平台集成了优化的光检测和测距-同时定位和绘图框架。该框架融合了来自光探测和测距(LiDAR)的点云数据、来自惯性测量单元(IMU)的姿态数据和全球导航卫星系统(GNSS)信号。因此,它可以准确地捕获非结构化果园中物体的点云。其次,利用迁移学习技术实现了小样本果园点云图的快速语义分割。最后,采用聚类分割算法分离柠檬树个体,提取树高、树中心位置、树冠大小等参数。本研究中提出的方法在一个非结构化柠檬园的真实场景中进行了评估。所提出的激光雷达SLAM框架显示出垂直均方根误差(RMSE)的显著降低。与FAST-LIO2和LIO-SAM相比,该框架的RMSE分别降低了68.55%和53.15%。基于迁移学习的分割实现了比基线PointNet++高3.6%的平均交联(mIoU)。此外,提取的树木参数与地面真值测量结果具有很强的相关性。这项工作为自动化果园感知提供了一个经过验证的端到端解决方案,为精确管理任务提供了关键数据。
{"title":"Research on real time mapping and transfer learning based 3D semantic segmentation of unstructured orchards","authors":"Sipei Li ,&nbsp;Sheng Wen ,&nbsp;Fagang Liu ,&nbsp;Haiyan Zhang ,&nbsp;Yufeng Ge ,&nbsp;Zhiyin Deng","doi":"10.1016/j.compag.2025.111323","DOIUrl":"10.1016/j.compag.2025.111323","url":null,"abstract":"<div><div>Precise semantic maps can provide considerable assistance in the intelligent management of unstructured orchards. However, successful extraction of useful tree information from complex environments to achieve autonomous unmanned aerial vehicle (UAV) operations in modern smart agriculture yet to be achieved.</div><div>Herein, we propose a real-time mapping and transfer learning-based 3D semantic segmentation framework to enable autonomous UAV operations in orchards. First, we constructed an autonomous perception platform. This platform integrates an optimized light detection and ranging–simultaneous localization and mapping framework. The framework fuses point cloud data from light detection and ranging (LiDAR), pose data from inertial measurement units (IMU), and Global Navigation satellite system (GNSS) signals. As a result, it can accurately capture point clouds of objects in unstructured orchards. Second, using transfer learning, we achieved rapid semantic segmentation of orchard point cloud maps with a small sample dataset. Finally, we employed a clustering segmentation algorithm to isolate individual lemon trees and extract parameters such as tree height, tree center location, and canopy size. The method proposed in this study was evaluated in a real-world scenario of an unstructured lemon orchard. The proposed LiDAR SLAM framework demonstrated a significant reduction in vertical root mean square error (RMSE). The framework achieved RMSE reductions of 68.55 % and 53.15 % against FAST-LIO2 and LIO-SAM, respectively. The transfer learning-based segmentation achieved a mean intersection over union (mIoU) 3.6 % higher than the baseline PointNet++. Furthermore, the extracted tree parameters exhibited a strong correlation with ground-truth measurements. This work provides a validated end-to-end solution for automated orchard perception, offering key data for precision management tasks.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"242 ","pages":"Article 111323"},"PeriodicalIF":8.9,"publicationDate":"2025-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145791868","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep learning–driven hyperspectral imaging for drought stress detection in dragoon lettuce for space production 用于空间生产的龙莴苣干旱胁迫检测的深度学习驱动高光谱成像
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-18 DOI: 10.1016/j.compag.2025.111321
Hangi Kim , Eun-Sung Park , Moon S. Kim , Insuck Baek , Blake Costine , LaShelle E. Spencer , Aubrie O’Rourke , Hoonsoo Lee , Geonwoo Kim , Changyeun Mo , Byoung-Kwan Cho
Sustainable plant cultivation is critical for supporting long-duration space missions by ensuring reliable food production in extraterrestrial environments where resources are severely limited and growth systems operate in closed-loop conditions. With crew members managing multiple critical mission tasks and having minimal time for plant care, autonomous stress detection systems must provide reliable, interpretable diagnostics to enable rapid, informed decision-making for crop management. This study utilized a custom hyperspectral imaging (HSI) system designed for space applications to develop an AI-driven diagnostic framework. We propose a novel SAM-ViT-3PE architecture that uniquely combines sparse spectral band selection with 3D spatial-spectral patch embedding, preserving rich spatial-spectral information typically lost in conventional ROI-averaged approaches. A key temporal finding identified Day 3 After Treatment (DAT 3) as the critical threshold where drought stress signatures become distinctly detectable, with accuracy dramatically improving from 72.2% to 95.9%. By focusing analysis on data from DAT 3 onward, the SAM-ViT-3PE model achieved superior performance compared to traditional ML methods and standard deep learning approaches, with accuracy of 95.4%, precision of 96.6% and recall of 94.1%. Furthermore, Explainable AI using Integrated Gradients enabled interpretable diagnostics through physiologically meaningful spectral bands and spatial stress patterns. These results demonstrate that the AI-enhanced HSI framework provides both high-accuracy autonomous detection and scientifically grounded interpretability essential for trustworthy crop management in resource-constrained space environments.
在资源严重有限、生长系统在闭环条件下运行的地外环境中,通过确保可靠的粮食生产,可持续植物种植对于支持长期空间任务至关重要。由于机组人员要管理多个关键任务,并且只有最少的时间来照顾植物,因此自主压力检测系统必须提供可靠、可解释的诊断,以便为作物管理提供快速、明智的决策。本研究利用为空间应用设计的定制高光谱成像(HSI)系统来开发人工智能驱动的诊断框架。我们提出了一种新的SAM-ViT-3PE架构,该架构独特地将稀疏频谱带选择与3D空间光谱补丁嵌入相结合,保留了传统roi平均方法中通常丢失的丰富空间光谱信息。一个关键的时间发现确定了处理后第3天(DAT 3)是干旱胁迫特征明显可检测的关键阈值,准确度从72.2%显著提高到95.9%。通过对data3以后的数据进行重点分析,sam - viti - 3pe模型与传统的ML方法和标准深度学习方法相比,取得了更好的性能,准确率为95.4%,精密度为96.6%,召回率为94.1%。此外,使用集成梯度的可解释人工智能通过生理上有意义的光谱带和空间应力模式实现了可解释的诊断。这些结果表明,人工智能增强的HSI框架提供了高精度的自主检测和科学的可解释性,对于资源受限的空间环境中可靠的作物管理至关重要。
{"title":"Deep learning–driven hyperspectral imaging for drought stress detection in dragoon lettuce for space production","authors":"Hangi Kim ,&nbsp;Eun-Sung Park ,&nbsp;Moon S. Kim ,&nbsp;Insuck Baek ,&nbsp;Blake Costine ,&nbsp;LaShelle E. Spencer ,&nbsp;Aubrie O’Rourke ,&nbsp;Hoonsoo Lee ,&nbsp;Geonwoo Kim ,&nbsp;Changyeun Mo ,&nbsp;Byoung-Kwan Cho","doi":"10.1016/j.compag.2025.111321","DOIUrl":"10.1016/j.compag.2025.111321","url":null,"abstract":"<div><div>Sustainable plant cultivation is critical for supporting long-duration space missions by ensuring reliable food production in extraterrestrial environments where resources are severely limited and growth systems operate in closed-loop conditions. With crew members managing multiple critical mission tasks and having minimal time for plant care, autonomous stress detection systems must provide reliable, interpretable diagnostics to enable rapid, informed decision-making for crop management. This study utilized a custom hyperspectral imaging (HSI) system designed for space applications to develop an AI-driven diagnostic framework. We propose a novel SAM-ViT-3PE architecture that uniquely combines sparse spectral band selection with 3D spatial-spectral patch embedding, preserving rich spatial-spectral information typically lost in conventional ROI-averaged approaches. A key temporal finding identified Day 3 After Treatment (DAT 3) as the critical threshold where drought stress signatures become distinctly detectable, with accuracy dramatically improving from 72.2% to 95.9%. By focusing analysis on data from DAT 3 onward, the SAM-ViT-3PE model achieved superior performance compared to traditional ML methods and standard deep learning approaches, with accuracy of 95.4%, precision of 96.6% and recall of 94.1%. Furthermore, Explainable AI using Integrated Gradients enabled interpretable diagnostics through physiologically meaningful spectral bands and spatial stress patterns. These results demonstrate that the AI-enhanced HSI framework provides both high-accuracy autonomous detection and scientifically grounded interpretability essential for trustworthy crop management in resource-constrained space environments.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"242 ","pages":"Article 111321"},"PeriodicalIF":8.9,"publicationDate":"2025-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145791791","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Low–cost IoT–enabled organic pH control system for smart aquaponic production in urban agriculture 城市农业智能水培生产的低成本物联网有机pH控制系统
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-18 DOI: 10.1016/j.compag.2025.111255
Ananta Sinchai , Chitsanucha Ardkaew , Tontrakan Jaiyen
This study presents a smart aquaponic system integrating low–cost IoT monitoring, automated organic pH control using food–grade orange juice, and computer vision–based plant growth assessment to enhance sustainable urban lettuce production. Two experimental rounds compared uncontrolled pH conditions with active regulation across 15 plants per group, with sample size adequacy confirmed through power analysis for large observed effects. The system maintained optimal pH (5.8–7.0) through automated feedback control of orange juice within a defined range, producing 13–30 % higher growth rates, supported by t-tests, effect sizes, confidence intervals, and regression analysis. Performance and cost efficiency were assessed against established control strategies, showing competitive outcomes under resource–limited conditions. Economic evaluation indicated a payback period of 9–59  months depending on local market factors. Overall, results demonstrate the feasibility of combining natural pH buffers with IoT feedback to reduce chemical inputs, sustain growth improvements, and enable scalable adoption in small– to medium–scale aquaponics.
本研究提出了一种集成低成本物联网监测、使用食品级橙汁自动有机pH控制和基于计算机视觉的植物生长评估的智能水培系统,以提高城市生菜的可持续生产。两轮实验将不受控制的pH条件与每组15个植物的主动调节进行了比较,通过功率分析证实了样本量的充足性。在t检验、效应量、置信区间和回归分析的支持下,该系统通过对橙汁的自动反馈控制,在限定范围内保持最佳pH值(5.8-7.0),生长期提高13 - 30%。根据既定的控制策略评估了性能和成本效率,显示了资源有限条件下的竞争结果。经济评价表明,根据当地市场因素,投资回收期为9-59个月。总体而言,研究结果证明了将天然pH缓冲液与物联网反馈相结合的可行性,以减少化学投入,维持生长改善,并使中小型鱼共生系统能够大规模采用。
{"title":"Low–cost IoT–enabled organic pH control system for smart aquaponic production in urban agriculture","authors":"Ananta Sinchai ,&nbsp;Chitsanucha Ardkaew ,&nbsp;Tontrakan Jaiyen","doi":"10.1016/j.compag.2025.111255","DOIUrl":"10.1016/j.compag.2025.111255","url":null,"abstract":"<div><div>This study presents a smart aquaponic system integrating low–cost IoT monitoring, automated organic pH control using food–grade orange juice, and computer vision–based plant growth assessment to enhance sustainable urban lettuce production. Two experimental rounds compared uncontrolled pH conditions with active regulation across 15 plants per group, with sample size adequacy confirmed through power analysis for large observed effects. The system maintained optimal pH (5.8–7.0) through automated feedback control of orange juice within a defined range, producing 13–30 % higher growth rates, supported by t-tests, effect sizes, confidence intervals, and regression analysis. Performance and cost efficiency were assessed against established control strategies, showing competitive outcomes under resource–limited conditions. Economic evaluation indicated a payback period of 9–59  months depending on local market factors. Overall, results demonstrate the feasibility of combining natural pH buffers with IoT feedback to reduce chemical inputs, sustain growth improvements, and enable scalable adoption in small– to medium–scale aquaponics.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"242 ","pages":"Article 111255"},"PeriodicalIF":8.9,"publicationDate":"2025-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145791790","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Lightweight model for beef cattle behavior recognition from quadruped robot video in grassland pastures 基于四足机器人视频的草地牧场肉牛行为识别轻量级模型
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-17 DOI: 10.1016/j.compag.2025.111329
Peigang Wei , Wei Sun , Shanshan Cao , Fantao Kong
Accurate and rapid identification of typical cattle behaviors is fundamental to disease diagnosis, estrus monitoring, calving prediction, and health assessment. Existing machine vision approaches for behavior recognition in medium-to-large livestock (such as pigs, cattle, and sheep) are mainly tailored for indoor barn conditions. These methods perform poorly in outdoor grazing environments, where variable lighting, complex backgrounds, group clustering with occlusion, and motion blur pose substantial challenges. Our study proposes MASM-YOLO, a lightweight beef cattle behavior recognition model based on quadrupedal robots and edge computing. Using YOLOv11s as the baseline, we designed the Multi-Scale Focus and Extraction Network (MSFEN) to mitigate detection challenges caused by spatial scale differences and motion blur while enhancing cross-scale feature interaction. We further constructed the Adaptive Decomposition and Alignment Head (ADAH) to improve recognition accuracy in scenarios with clustering and occlusion. The lightweight feature extraction network StarNet optimizes the backbone structure, significantly reducing parameter count and computational load. The Inner-MPDIoU loss function is introduced to enhance the convergence and robustness of bounding box regression. Results demonstrate that MASM-YOLO achieves an [email protected] of 90.4 % on the beef cattle behavior test set, surpassing the baseline model by 1.9 percentage points and significantly outperforming mainstream CNN and Transformer models. The model contains 4.0 M parameters and requires 18.2G FLOPs, representing reductions of 57.4 % and 14.6 % compared to the baseline. After TensorRT optimization on the NVIDIA Jetson Orin NX edge platform, MASM-YOLO achieves real-time inference at 36FPS while maintaining an [email protected] of 89.6 %, validating its efficiency and feasibility on robotic platforms. This study demonstrates lightweight intelligent recognition of beef cattle behavior in grassland pastures through quadruped robot vision. Its uniqueness lies in constructing a new dataset and completing practical deployment verification on a robotic platform, providing valuable references and insights for automatic real-time livestock behavior recognition using ground-based mobile intelligent equipment in open grazing environments.
准确、快速地识别牛的典型行为是疾病诊断、发情监测、产犊预测和健康评估的基础。现有的用于大中型牲畜(如猪、牛、羊)行为识别的机器视觉方法主要针对室内畜棚条件量身定制。这些方法在户外放牧环境中表现不佳,在户外放牧环境中,可变光照、复杂背景、遮挡和运动模糊构成了实质性的挑战。本研究提出了一种基于四足机器人和边缘计算的轻量级肉牛行为识别模型MASM-YOLO。以yolov11为基准,设计了多尺度聚焦与提取网络(MSFEN),以缓解空间尺度差异和运动模糊带来的检测挑战,同时增强跨尺度特征交互。为了提高聚类和遮挡场景下的识别精度,我们进一步构建了自适应分解和对准头(ADAH)。轻量级特征提取网络StarNet优化了主干结构,显著减少了参数个数和计算量。为了提高边界盒回归的收敛性和鲁棒性,引入了Inner-MPDIoU损失函数。结果表明,MASM-YOLO在肉牛行为测试集上达到了90.4%的[email protected],超过基准模型1.9个百分点,显著优于主流的CNN和Transformer模型。该模型包含4.0 M参数,需要18.2G FLOPs,与基线相比分别降低了57.4%和14.6%。在NVIDIA Jetson Orin NX边缘平台上进行TensorRT优化后,MASM-YOLO实现了36FPS的实时推理,同时保持了89.6%的[email protected],验证了其在机器人平台上的效率和可行性。本研究利用四足机器人视觉实现了草地牧场肉牛行为的轻量级智能识别。其独特之处在于构建了新的数据集,并在机器人平台上完成了实际部署验证,为地面移动智能设备在露天放牧环境下自动实时识别牲畜行为提供了有价值的参考和见解。
{"title":"Lightweight model for beef cattle behavior recognition from quadruped robot video in grassland pastures","authors":"Peigang Wei ,&nbsp;Wei Sun ,&nbsp;Shanshan Cao ,&nbsp;Fantao Kong","doi":"10.1016/j.compag.2025.111329","DOIUrl":"10.1016/j.compag.2025.111329","url":null,"abstract":"<div><div>Accurate and rapid identification of typical cattle behaviors is fundamental to disease diagnosis, estrus monitoring, calving prediction, and health assessment. Existing machine vision approaches for behavior recognition in medium-to-large livestock (such as pigs, cattle, and sheep) are mainly tailored for indoor barn conditions. These methods perform poorly in outdoor grazing environments, where variable lighting, complex backgrounds, group clustering with occlusion, and motion blur pose substantial challenges. Our study proposes MASM-YOLO, a lightweight beef cattle behavior recognition model based on quadrupedal robots and edge computing. Using YOLOv11s as the baseline, we designed the Multi-Scale Focus and Extraction Network (MSFEN) to mitigate detection challenges caused by spatial scale differences and motion blur while enhancing cross-scale feature interaction. We further constructed the Adaptive Decomposition and Alignment Head (ADAH) to improve recognition accuracy in scenarios with clustering and occlusion. The lightweight feature extraction network StarNet optimizes the backbone structure, significantly reducing parameter count and computational load. The Inner-MPDIoU loss function is introduced to enhance the convergence and robustness of bounding box regression. Results demonstrate that MASM-YOLO achieves an [email protected] of 90.4 % on the beef cattle behavior test set, surpassing the baseline model by 1.9 percentage points and significantly outperforming mainstream CNN and Transformer models. The model contains 4.0 M parameters and requires 18.2G FLOPs, representing reductions of 57.4 % and 14.6 % compared to the baseline. After TensorRT optimization on the NVIDIA Jetson Orin NX edge platform, MASM-YOLO achieves real-time inference at 36FPS while maintaining an [email protected] of 89.6 %, validating its efficiency and feasibility on robotic platforms. This study demonstrates lightweight intelligent recognition of beef cattle behavior in grassland pastures through quadruped robot vision. Its uniqueness lies in constructing a new dataset and completing practical deployment verification on a robotic platform, providing valuable references and insights for automatic real-time livestock behavior recognition using ground-based mobile intelligent equipment in open grazing environments.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"242 ","pages":"Article 111329"},"PeriodicalIF":8.9,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145791793","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An efficient, scalable, and high-precision multifunctional intelligent navigation system for agricultural machinery 一种高效、可扩展、高精度的农业机械多功能智能导航系统
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-17 DOI: 10.1016/j.compag.2025.111336
Jianxing Xiao , Qiang Sheng , Yajun An , Ning Wang , Tianhai Wang , Shunda Li , Han Li , Man Zhang
Obstacle detection and automatic navigation are key components of intelligent navigation systems for agricultural machinery. Most current systems adopt an integrated design with limited flexibility and scalability. To address these limitations, this study proposes a loosely coupled intelligent navigation system based on an edge computing platform and an Microcontroller Unit (MCU), employing a distributed computing architecture for parallel processing of obstacle detection and navigation tasks. A method based on the fusion of 3D LiDAR and RGB camera is proposed to simultaneously identify obstacles category and distance accurately. In addition, an adaptive path tracking strategy based on the Pure Pursuit algorithm with Proportional Integral Derivative (PID) algorithm is implemented, which integrates lateral error from the Stanley controller and heading error, and incorporates a fuzzy logic-based gain adjustment to improve convergence under large initial deviations. Experimental results demonstrated that the system achieved real-time, high-precision obstacle detection and, compared with the Pure Pursuit method, improved alignment distance by at least 24.8 % and alignment time by at least 22.6 %, indicating strong flexibility, scalability, and stability for both obstacle detection and path tracking under realistic agricultural conditions.
障碍物检测和自动导航是农业机械智能导航系统的关键组成部分。目前大多数系统采用集成设计,灵活性和可扩展性有限。为了解决这些限制,本研究提出了一种基于边缘计算平台和微控制器单元(MCU)的松耦合智能导航系统,采用分布式计算架构并行处理障碍物检测和导航任务。提出了一种基于三维激光雷达和RGB相机融合的障碍物分类和距离识别方法。此外,基于比例积分导数(PID)算法的纯追踪算法实现了自适应路径跟踪策略,该策略集成了Stanley控制器的横向误差和航向误差,并结合基于模糊逻辑的增益调整,提高了大初始偏差下的收敛性。实验结果表明,该系统实现了实时、高精度的障碍物检测,与Pure Pursuit方法相比,对准距离至少提高了24.8%,对准时间至少提高了22.6%,表明该系统在现实农业条件下的障碍物检测和路径跟踪具有很强的灵活性、可扩展性和稳定性。
{"title":"An efficient, scalable, and high-precision multifunctional intelligent navigation system for agricultural machinery","authors":"Jianxing Xiao ,&nbsp;Qiang Sheng ,&nbsp;Yajun An ,&nbsp;Ning Wang ,&nbsp;Tianhai Wang ,&nbsp;Shunda Li ,&nbsp;Han Li ,&nbsp;Man Zhang","doi":"10.1016/j.compag.2025.111336","DOIUrl":"10.1016/j.compag.2025.111336","url":null,"abstract":"<div><div>Obstacle detection and automatic navigation are key components of intelligent navigation systems for agricultural machinery. Most current systems adopt an integrated design with limited flexibility and scalability. To address these limitations, this study proposes a loosely coupled intelligent navigation system based on an edge computing platform and an Microcontroller Unit (MCU), employing a distributed computing architecture for parallel processing of obstacle detection and navigation tasks. A method based on the fusion of 3D LiDAR and RGB camera is proposed to simultaneously identify obstacles category and distance accurately. In addition, an adaptive path tracking strategy based on the Pure Pursuit algorithm with Proportional Integral Derivative (PID) algorithm is implemented, which integrates lateral error from the Stanley controller and heading error, and incorporates a fuzzy logic-based gain adjustment to improve convergence under large initial deviations. Experimental results demonstrated that the system achieved real-time, high-precision obstacle detection and, compared with the Pure Pursuit method, improved alignment distance by at least 24.8 % and alignment time by at least 22.6 %, indicating strong flexibility, scalability, and stability for both obstacle detection and path tracking under realistic agricultural conditions.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"242 ","pages":"Article 111336"},"PeriodicalIF":8.9,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145791864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computers and Electronics in Agriculture
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1