首页 > 最新文献

Artificial Intelligence in Agriculture最新文献

英文 中文
Integrating 3D detection networks and dynamic temporal phenotyping for wheat yield classification and prediction 结合三维检测网络和动态时间表型进行小麦产量分类和预测
IF 12.4 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-06 DOI: 10.1016/j.aiia.2025.12.001
Honghao Zhou , Bingxi Qin , Qing Li , Wenlong Su , Shaowei Liang , Haijiang Min , Jingrong Zang , Shichao Jin , Dong Jiang , Jiawei Chen
Automated phenotyping of wheat growth stages from 3D point clouds is still limited. The study presents a concise framework that reconstructs multi-view UAS imagery into 3D point clouds (jointing to maturity) and performs plot-level phenotyping. A novel 3D wheat plot detection network—integrating spatial–channel coordinated attention and area attention modules—improves depth-direction feature recognition, and a point-cloud-density-based row segmentation algorithm enables planting-row-scale plot delineation. A supporting software system facilitates 3D visualization and automated extraction of phenotypic parameters. We introduce a dynamic phenotypic index of five temporal metrics (growth stage, slow growth stage, height/area reduction stage, maximum height/area difference stage, and height/area change rate) for growth-stage classification and yield prediction using static and time-series models. Experiments show strong agreement between predicted and measured plot heights (R2 = 0.937); the detection net achieved AP3D = 94.15 % and APBEV = 95.35 % in “easy” mode; and a Bi-LSTM incorporating dynamic traits reached 82.37 % prediction accuracy for leaf area and yield, a 6.14 % improvement over static-trait models. This workflow supports high-throughput 3D phenotyping and reliable yield estimation for precision agriculture.
小麦生长阶段的三维点云自动表型分析仍然有限。该研究提出了一个简洁的框架,将多视图UAS图像重建为3D点云(连接到成熟度),并执行情节水平表型。基于空间通道协调注意和区域注意模块的小麦三维地块检测网络改进了深度-方向特征识别,基于点云密度的行分割算法实现了种植行尺度的地块划分。一个支持的软件系统促进了表型参数的3D可视化和自动提取。采用静态和时间序列模型,引入生长期、慢生长期、高/面积减少期、最大高/面积差期和高/面积变化率五个时间指标的动态表型指数,用于生长期分类和产量预测。实验结果表明,预测值与实测值吻合较好(R2 = 0.937);“easy”模式下检测网AP3D = 94.15%, APBEV = 95.35%;结合动态性状的Bi-LSTM对叶面积和产量的预测准确率达到82.37%,比静态性状模型提高6.14%。该工作流程支持高通量3D表型和可靠的精准农业产量估计。
{"title":"Integrating 3D detection networks and dynamic temporal phenotyping for wheat yield classification and prediction","authors":"Honghao Zhou ,&nbsp;Bingxi Qin ,&nbsp;Qing Li ,&nbsp;Wenlong Su ,&nbsp;Shaowei Liang ,&nbsp;Haijiang Min ,&nbsp;Jingrong Zang ,&nbsp;Shichao Jin ,&nbsp;Dong Jiang ,&nbsp;Jiawei Chen","doi":"10.1016/j.aiia.2025.12.001","DOIUrl":"10.1016/j.aiia.2025.12.001","url":null,"abstract":"<div><div>Automated phenotyping of wheat growth stages from 3D point clouds is still limited. The study presents a concise framework that reconstructs multi-view UAS imagery into 3D point clouds (jointing to maturity) and performs plot-level phenotyping. A novel 3D wheat plot detection network—integrating spatial–channel coordinated attention and area attention modules—improves depth-direction feature recognition, and a point-cloud-density-based row segmentation algorithm enables planting-row-scale plot delineation. A supporting software system facilitates 3D visualization and automated extraction of phenotypic parameters. We introduce a dynamic phenotypic index of five temporal metrics (growth stage, slow growth stage, height/area reduction stage, maximum height/area difference stage, and height/area change rate) for growth-stage classification and yield prediction using static and time-series models. Experiments show strong agreement between predicted and measured plot heights (R<sup>2</sup> = 0.937); the detection net achieved AP<sub>3D</sub> = 94.15 % and AP<sub>BEV</sub> = 95.35 % in “easy” mode; and a Bi-LSTM incorporating dynamic traits reached 82.37 % prediction accuracy for leaf area and yield, a 6.14 % improvement over static-trait models. This workflow supports high-throughput 3D phenotyping and reliable yield estimation for precision agriculture.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"16 1","pages":"Pages 603-618"},"PeriodicalIF":12.4,"publicationDate":"2025-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145747432","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimizing herbicide reduction: A simulation approach using Artificial Intelligence and different nozzle configurations 优化除草剂减少:使用人工智能和不同喷嘴配置的模拟方法
IF 12.4 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-01 DOI: 10.1016/j.aiia.2025.11.011
Renato H. Furlanetto , Ana C. Buzanini , Arnold W. Schumann , Nathan S. Boyd
Targeted application aims to minimize product usage by spraying only where needed. However, there is a lack of tools to evaluate the potential savings and the model's limitations. To address this, we developed a computer simulation using two YOLOv8x models (bounding boxes and segmentation masks) to assess spray volume reduction (SVR) and nozzle limitations across varying weed densities. We used Python, OpenCV, and Tkinter to analyze videos from row-middle trials at the University of Florida. The system tracks weeds crossing a horizontal detection line within a 70 cm row-middle width, dividing the frame into one to eight vertical zones representing nozzle distribution. When a weed is detected, the corresponding frame is saved. The system calculates the activation time in seconds by considering the total number of video frames. Spray volume calculations were based on manual measurements of a TeeJet 8001VS nozzle tip, which dispenses a known volume of liquid per second at 35 PSI. The activation time was multiplied by this rate to estimate the targeted spray volume. The broadcast application volume was calculated by multiplying the total video duration by the same tip output. The results showed that the models achieved up to 96 % accuracy (mAP) with no statistical difference. A polynomial model for low and medium weed densities demonstrated SVR of 74 % (six nozzles) and 57 % (seven nozzles). A linear model for high density achieved a 40 % reduction. The lowest reduction occurred with a single nozzle (4 % for medium, 2 % for high density). These findings demonstrate that nozzle density significantly impacted spray reduction at medium and high densities while low-density savings remained consistent.
定向应用的目的是通过只在需要的地方喷洒来减少产品的使用。然而,缺乏工具来评估潜在的节省和模型的局限性。为了解决这个问题,我们开发了一个使用两个YOLOv8x模型(边界盒和分割掩模)的计算机模拟,以评估不同杂草密度下的喷雾体积减少(SVR)和喷嘴限制。我们使用Python、OpenCV和Tkinter来分析佛罗里达大学(University of Florida)中排试验的视频。该系统在一个70厘米宽的中线范围内跟踪穿过水平检测线的杂草,将框架划分为代表喷嘴分布的一到八个垂直区域。当检测到杂草时,保存相应的帧。系统通过考虑视频帧总数来计算激活时间(以秒为单位)。喷雾体积的计算是基于TeeJet 8001VS喷嘴尖端的手动测量,该喷嘴在35psi下每秒分配已知体积的液体。激活时间乘以这个速率来估计目标喷雾体积。通过将总视频持续时间乘以相同的提示输出来计算广播应用音量。结果表明,模型的准确率高达96%,无统计学差异。低、中杂草密度的多项式模型显示SVR分别为74%(6个喷嘴)和57%(7个喷嘴)。高密度的线性模型实现了40%的减少。单喷嘴的降低率最低(中等密度为4%,高密度为2%)。这些研究结果表明,在中高密度和低密度情况下,喷嘴密度显著影响了喷雾减少,而低密度情况下,喷嘴密度保持不变。
{"title":"Optimizing herbicide reduction: A simulation approach using Artificial Intelligence and different nozzle configurations","authors":"Renato H. Furlanetto ,&nbsp;Ana C. Buzanini ,&nbsp;Arnold W. Schumann ,&nbsp;Nathan S. Boyd","doi":"10.1016/j.aiia.2025.11.011","DOIUrl":"10.1016/j.aiia.2025.11.011","url":null,"abstract":"<div><div>Targeted application aims to minimize product usage by spraying only where needed. However, there is a lack of tools to evaluate the potential savings and the model's limitations. To address this, we developed a computer simulation using two YOLOv8x models (bounding boxes and segmentation masks) to assess spray volume reduction (SVR) and nozzle limitations across varying weed densities. We used Python, OpenCV, and Tkinter to analyze videos from row-middle trials at the University of Florida. The system tracks weeds crossing a horizontal detection line within a 70 cm row-middle width, dividing the frame into one to eight vertical zones representing nozzle distribution. When a weed is detected, the corresponding frame is saved. The system calculates the activation time in seconds by considering the total number of video frames. Spray volume calculations were based on manual measurements of a TeeJet 8001VS nozzle tip, which dispenses a known volume of liquid per second at 35 PSI. The activation time was multiplied by this rate to estimate the targeted spray volume. The broadcast application volume was calculated by multiplying the total video duration by the same tip output. The results showed that the models achieved up to 96 % accuracy (mAP) with no statistical difference. A polynomial model for low and medium weed densities demonstrated SVR of 74 % (six nozzles) and 57 % (seven nozzles). A linear model for high density achieved a 40 % reduction. The lowest reduction occurred with a single nozzle (4 % for medium, 2 % for high density). These findings demonstrate that nozzle density significantly impacted spray reduction at medium and high densities while low-density savings remained consistent.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"16 1","pages":"Pages 592-602"},"PeriodicalIF":12.4,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145747431","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SCLFormer: A synergistic convolution-linear attention transformer for hyperspectral image classification of mechanical damage in maize kernels SCLFormer:一种用于玉米籽粒机械损伤高光谱图像分类的协同卷积-线性注意转换器
IF 12.4 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-11-27 DOI: 10.1016/j.aiia.2025.11.010
Yiqiang Zheng , Jun Fu , Fengshuang Liu , Haiming Zhao , Jindai Liu
Classifying mechanical damage in maize kernels using hyperspectral imaging is crucial for food security and loss reduction. Existing methods are constrained by high computational complexity and limited precision in detecting subtle damages, such as pericarp damage and kernel cracks. To address these challenges, we introduce a novel algorithm, the Synergistic Convolution and Linear Attention Transformer (SCLFormer). By replacing traditional softmax attention with linear attention, we reduce computational complexity from quadratic to linear, thereby enhancing efficiency. Integrating convolutional operations into the encoder enriches local prior information for global feature modeling, improving classification accuracy. SCLFormer achieves an overall accuracy of 97.08 % in classifying maize kernel damage, with over 85 % accuracy for cracked kernel and pericarp damage. Compared to softmax attention, SCLFormer reduces training and testing times by 355.27 s (16.86 %) and 0.71 s (23.67 %), respectively. Additionally, we propose a modular hyperspectral image-level classification framework that can integrate existing pixel-level feature extraction networks to achieve classification accuracies exceeding 80 %, demonstrating the framework's scalability. SCLFormer, serving as the framework's dedicated feature extraction component, provides a robust solution for maize kernel damage classification and exhibits substantial potential for broader spatial-scale applications. This framework establishes a novel technical paradigm for hyperspectral image-wise classification of other agricultural products.
利用高光谱成像技术对玉米籽粒机械损伤进行分类对粮食安全和减少损失至关重要。现有方法在检测果皮损伤和果仁裂纹等细微损伤时,计算量大、精度低。为了解决这些挑战,我们引入了一种新的算法,即协同卷积和线性注意力转换器(SCLFormer)。通过将传统的softmax注意力替换为线性注意力,将计算复杂度从二次型降低到线性型,从而提高了效率。将卷积运算集成到编码器中,丰富了全局特征建模的局部先验信息,提高了分类精度。SCLFormer对玉米籽粒损伤分类的总体准确率为97.08%,对裂粒和果皮损伤分类的准确率超过85%。与softmax相比,SCLFormer的训练和测试时间分别减少了355.27 s(16.86%)和0.71 s(23.67%)。此外,我们提出了一个模块化的高光谱图像级分类框架,该框架可以集成现有的像素级特征提取网络,实现超过80%的分类精度,证明了框架的可扩展性。SCLFormer作为该框架的专用特征提取组件,为玉米籽粒损伤分类提供了一个强大的解决方案,并显示出更广泛的空间尺度应用潜力。该框架为其他农产品的高光谱图像分类建立了一种新的技术范式。
{"title":"SCLFormer: A synergistic convolution-linear attention transformer for hyperspectral image classification of mechanical damage in maize kernels","authors":"Yiqiang Zheng ,&nbsp;Jun Fu ,&nbsp;Fengshuang Liu ,&nbsp;Haiming Zhao ,&nbsp;Jindai Liu","doi":"10.1016/j.aiia.2025.11.010","DOIUrl":"10.1016/j.aiia.2025.11.010","url":null,"abstract":"<div><div>Classifying mechanical damage in maize kernels using hyperspectral imaging is crucial for food security and loss reduction. Existing methods are constrained by high computational complexity and limited precision in detecting subtle damages, such as pericarp damage and kernel cracks. To address these challenges, we introduce a novel algorithm, the Synergistic Convolution and Linear Attention Transformer (SCLFormer). By replacing traditional softmax attention with linear attention, we reduce computational complexity from quadratic to linear, thereby enhancing efficiency. Integrating convolutional operations into the encoder enriches local prior information for global feature modeling, improving classification accuracy. SCLFormer achieves an overall accuracy of 97.08 % in classifying maize kernel damage, with over 85 % accuracy for cracked kernel and pericarp damage. Compared to softmax attention, SCLFormer reduces training and testing times by 355.27 s (16.86 %) and 0.71 s (23.67 %), respectively. Additionally, we propose a modular hyperspectral image-level classification framework that can integrate existing pixel-level feature extraction networks to achieve classification accuracies exceeding 80 %, demonstrating the framework's scalability. SCLFormer, serving as the framework's dedicated feature extraction component, provides a robust solution for maize kernel damage classification and exhibits substantial potential for broader spatial-scale applications. This framework establishes a novel technical paradigm for hyperspectral image-wise classification of other agricultural products.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"16 1","pages":"Pages 565-577"},"PeriodicalIF":12.4,"publicationDate":"2025-11-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145693322","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-arm robotic system and strategy for the automatic packaging of apples 苹果自动包装的多臂机器人系统及策略
IF 12.4 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-11-26 DOI: 10.1016/j.aiia.2025.11.006
Yizhi Zhang , Liping Chen , Xin Li , Qicheng Li , Jiangbo Li
Packaging is a crucial step in the commercial distribution of apples after harvest. However, achieving fast and accurate automated packaging remains a challenge. This study proposed for the first time a highly integrated four-arm robotic system and an effective implementation strategy specifically designed for automatic packaging of apples. In order to achieve accurate and simultaneous detection of multi-channel apples in complex visual conditions, an enhanced target detection model (i.e. YOLOv8s-GUC) based on the YOLOv8s architecture was developed by combining with a global attention mechanism (GAM), UniRepLKNet large kernel convolution structure, and CARAFE lightweight upsampling module. The results demonstrated that the YOLOv8s-GUC model can significantly improve detection accuracy and generalization for apples and related small features (such as stems and calyxes), achieving an [email protected] of 99.5 %. Furthermore, considering the complexity and real-time constraints of task planning in multi-arm robotic operations, this study further proposed an intelligent scheduling algorithm based on a Deep Q-Network (DQN), which enables efficient collaboration and collision-free operation among the robotic arms through real-time state perception and online decision-making. The results of simulations and real-world experiments indicated that the developed multi-arm robotic packaging system and scheduling strategy had high operational stability and efficiency in apple packaging, with a success rate of 100 % and an average packaging time of less than one second per apple. This study provides an effective and reliable solution for automated apple packaging.
包装是苹果收获后商业分销的关键一步。然而,实现快速和准确的自动化包装仍然是一个挑战。本研究首次提出了一种专门为苹果自动包装设计的高度集成的四臂机器人系统和有效的实施策略。为了实现复杂视觉条件下多通道苹果的准确、同时检测,结合全局注意机制(GAM)、UniRepLKNet大核卷积结构和CARAFE轻量级上采样模块,开发了基于YOLOv8s架构的增强型目标检测模型(即YOLOv8s- guc)。结果表明,YOLOv8s-GUC模型可以显著提高苹果及其相关小特征(如茎和花萼)的检测精度和泛化,达到99.5%的[email protected]。此外,考虑到多臂机器人作业任务规划的复杂性和实时性约束,本研究进一步提出了一种基于深度q网络(Deep Q-Network, DQN)的智能调度算法,通过实时状态感知和在线决策,实现机械臂间高效协同、无碰撞作业。仿真和实际实验结果表明,所开发的多臂机器人包装系统和调度策略在苹果包装中具有较高的运行稳定性和效率,成功率为100%,平均每个苹果的包装时间小于1秒。本研究为苹果自动化包装提供了有效可靠的解决方案。
{"title":"Multi-arm robotic system and strategy for the automatic packaging of apples","authors":"Yizhi Zhang ,&nbsp;Liping Chen ,&nbsp;Xin Li ,&nbsp;Qicheng Li ,&nbsp;Jiangbo Li","doi":"10.1016/j.aiia.2025.11.006","DOIUrl":"10.1016/j.aiia.2025.11.006","url":null,"abstract":"<div><div>Packaging is a crucial step in the commercial distribution of apples after harvest. However, achieving fast and accurate automated packaging remains a challenge. This study proposed for the first time a highly integrated four-arm robotic system and an effective implementation strategy specifically designed for automatic packaging of apples. In order to achieve accurate and simultaneous detection of multi-channel apples in complex visual conditions, an enhanced target detection model (i.e. YOLOv8s-GUC) based on the YOLOv8s architecture was developed by combining with a global attention mechanism (GAM), UniRepLKNet large kernel convolution structure, and CARAFE lightweight upsampling module. The results demonstrated that the YOLOv8s-GUC model can significantly improve detection accuracy and generalization for apples and related small features (such as stems and calyxes), achieving an [email protected] of 99.5 %. Furthermore, considering the complexity and real-time constraints of task planning in multi-arm robotic operations, this study further proposed an intelligent scheduling algorithm based on a Deep Q-Network (DQN), which enables efficient collaboration and collision-free operation among the robotic arms through real-time state perception and online decision-making. The results of simulations and real-world experiments indicated that the developed multi-arm robotic packaging system and scheduling strategy had high operational stability and efficiency in apple packaging, with a success rate of 100 % and an average packaging time of less than one second per apple. This study provides an effective and reliable solution for automated apple packaging.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"16 1","pages":"Pages 578-591"},"PeriodicalIF":12.4,"publicationDate":"2025-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145693435","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PlantSegNeRF: A few-shot, cross-species method for plant 3D instance point cloud reconstruction via joint-channel NeRF with multi-view image instance matching PlantSegNeRF:一种基于联合通道NeRF和多视图图像实例匹配的植物三维实例点云重建方法
IF 12.4 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-11-24 DOI: 10.1016/j.aiia.2025.11.009
Xin Yang , Ruiming Du , Hanyang Huang , Jiayang Xie , Pengyao Xie , Leisen Fang , Ziyue Guo , Nanjun Jiang , Yu Jiang , Haiyan Cen
Organ segmentation of plant point clouds is a prerequisite for the high-resolution and accurate extraction of organ-level phenotypic traits. Although the fast development of deep learning has boosted much research on segmentation of plant point clouds, the existing techniques for organ segmentation still face limitations in resolution, segmentation accuracy, and generalizability across various plant species. In this study, we proposed a novel approach called plant segmentation neural radiance fields (PlantSegNeRF), aiming to directly generate high-precision instance point clouds from multi-view RGB image sequences for a wide range of plant species. PlantSegNeRF performed two-dimensional (2D) instance segmentation on the multi-view images to generate instance masks for each organ with a corresponding instance identification (ID). The multi-view instance IDs corresponding to the same plant organ were then matched and refined using a specially designed instance matching (IM) module. The instance NeRF was developed to render an implicit scene containing color, density, semantic and instance information, which was ultimately converted into high-precision plant instance point clouds based on volume density. The results proved that in semantic segmentation of point clouds, PlantSegNeRF outperformed the commonly used methods, demonstrating an average improvement of 16.1 %, 18.3 %, 17.8 %, and 24.2 % in precision, recall, F1-score, and intersection over union (IoU) compared to the second-best results on structurally complex datasets. More importantly, PlantSegNeRF exhibited significant advantages in instance segmentation. Across all plant datasets, it achieved average improvements of 11.7 %, 38.2 %, 32.2 % and 25.3 % in mean precision (mPrec), mean recall (mRec), mean coverage (mCov), and mean weighted coverage (mWCov), respectively. Furthermore, PlantSegNeRF demonstrates superior few-shot, cross-species performance, requiring only multi-view images of few plants to train models applicable to specific or similar varieties. This study extends organ-level plant phenotyping and provides a high-throughput way to supply high-quality 3D data for developing large-scale artificial intelligence (AI) models in plant science.
植物点云的器官分割是高分辨率和准确提取器官水平表型性状的先决条件。尽管深度学习的快速发展促进了植物点云分割的研究,但现有的器官分割技术在分辨率、分割精度和跨植物物种的可泛化性等方面仍然存在局限性。在这项研究中,我们提出了一种新的方法,称为植物分割神经辐射场(PlantSegNeRF),旨在从多视角RGB图像序列中直接生成高精度的实例点云,用于广泛的植物物种。PlantSegNeRF对多视图图像进行二维(2D)实例分割,为每个器官生成具有相应实例标识(ID)的实例掩码。然后使用专门设计的实例匹配(IM)模块对同一植物器官对应的多视图实例id进行匹配和细化。开发实例NeRF来渲染包含颜色、密度、语义和实例信息的隐式场景,最终根据体积密度转换成高精度的植物实例点云。结果表明,在点云的语义分割中,PlantSegNeRF优于常用的方法,在精度、召回率、f1得分和交汇比(IoU)方面,与结构复杂数据集的第二好的结果相比,平均提高了16.1%、18.3%、17.8%和24.2%。更重要的是,PlantSegNeRF在实例分割方面表现出显著的优势。在所有植物数据集中,它在平均精度(mPrec)、平均召回率(mRec)、平均覆盖率(mCov)和平均加权覆盖率(mWCov)方面分别实现了11.7%、38.2%、32.2%和25.3%的平均改进。此外,PlantSegNeRF展示了优越的少拍、跨物种性能,只需要少数植物的多视图图像来训练适用于特定或类似品种的模型。该研究扩展了器官水平的植物表型,为植物科学中大规模人工智能(AI)模型的开发提供了高通量的高质量3D数据。
{"title":"PlantSegNeRF: A few-shot, cross-species method for plant 3D instance point cloud reconstruction via joint-channel NeRF with multi-view image instance matching","authors":"Xin Yang ,&nbsp;Ruiming Du ,&nbsp;Hanyang Huang ,&nbsp;Jiayang Xie ,&nbsp;Pengyao Xie ,&nbsp;Leisen Fang ,&nbsp;Ziyue Guo ,&nbsp;Nanjun Jiang ,&nbsp;Yu Jiang ,&nbsp;Haiyan Cen","doi":"10.1016/j.aiia.2025.11.009","DOIUrl":"10.1016/j.aiia.2025.11.009","url":null,"abstract":"<div><div>Organ segmentation of plant point clouds is a prerequisite for the high-resolution and accurate extraction of organ-level phenotypic traits. Although the fast development of deep learning has boosted much research on segmentation of plant point clouds, the existing techniques for organ segmentation still face limitations in resolution, segmentation accuracy, and generalizability across various plant species. In this study, we proposed a novel approach called plant segmentation neural radiance fields (PlantSegNeRF), aiming to directly generate high-precision instance point clouds from multi-view RGB image sequences for a wide range of plant species. PlantSegNeRF performed two-dimensional (2D) instance segmentation on the multi-view images to generate instance masks for each organ with a corresponding instance identification (ID). The multi-view instance IDs corresponding to the same plant organ were then matched and refined using a specially designed instance matching (IM) module. The instance NeRF was developed to render an implicit scene containing color, density, semantic and instance information, which was ultimately converted into high-precision plant instance point clouds based on volume density. The results proved that in semantic segmentation of point clouds, PlantSegNeRF outperformed the commonly used methods, demonstrating an average improvement of 16.1 %, 18.3 %, 17.8 %, and 24.2 % in precision, recall, F1-score, and intersection over union (IoU) compared to the second-best results on structurally complex datasets. More importantly, PlantSegNeRF exhibited significant advantages in instance segmentation. Across all plant datasets, it achieved average improvements of 11.7 %, 38.2 %, 32.2 % and 25.3 % in mean precision (mPrec), mean recall (mRec), mean coverage (mCov), and mean weighted coverage (mWCov), respectively. Furthermore, PlantSegNeRF demonstrates superior few-shot, cross-species performance, requiring only multi-view images of few plants to train models applicable to specific or similar varieties. This study extends organ-level plant phenotyping and provides a high-throughput way to supply high-quality 3D data for developing large-scale artificial intelligence (AI) models in plant science.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"16 1","pages":"Pages 546-564"},"PeriodicalIF":12.4,"publicationDate":"2025-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145623318","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-scale feature alignment network for 19-class semantic segmentation in agricultural environments 农业环境下19类语义分割的多尺度特征对齐网络
IF 12.4 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-11-22 DOI: 10.1016/j.aiia.2025.11.008
Zhi-xin Yao , Hao Wang , Zhi-jun Meng , Liang-liang Yang , Tai-hong Zhang
To improve environmental perception and ensure reliable agricultural machinery navigation during field transitions under unstructured farm road conditions, this study utilizes high-resolution RGB camera vision navigation technology to propose a Multi-Scale Feature Alignment Network (MSFA-Net) for 19-class semantic segmentation of agricultural environment, which includes information such as roads, pedestrians, and vehicles. MSFA-Net introduces two key innovations: the DASP module, which integrates multi-scale feature extraction with dual attention mechanisms (spatial and channel), and the MSFA architecture, which enables robust boundary extraction and mitigates interference from lighting variations and obstacles like vegetation. Compared to existing models, MSFA-Net uniquely combines efficient multi-scale feature extraction with real-time inference capabilities, achieving an mIoU of 84.46 % and an mPA of 96.10 %. For 512 × 512 input images, the model processes an average of 26 images/s on a GTX 1650Ti, with a boundary extraction error of less than 0.47 m within 20 m. These results indicate that the proposed MSFA-Net can significantly reduce navigation errors and improve the perception stability of agricultural machinery during field operations. Furthermore, the model can be exported to ONNX or TensorFlow Lite formats, facilitating efficient deployment on embedded devices and existing farm navigation systems.
为了提高农业机械在非结构化农田道路条件下的环境感知能力,保证农机导航的可靠性,本研究利用高分辨率RGB相机视觉导航技术,提出了一种多尺度特征对齐网络(MSFA-Net),对包括道路、行人、车辆等信息在内的农业环境进行19类语义分割。MSFA- net引入了两个关键的创新:DASP模块,它集成了具有双重注意机制(空间和通道)的多尺度特征提取;MSFA架构,它可以实现鲁棒的边界提取,并减轻光照变化和植被等障碍物的干扰。与现有模型相比,MSFA-Net独特地将高效的多尺度特征提取与实时推理能力相结合,实现了84.46%的mIoU和96.10%的mPA。对于512 × 512的输入图像,该模型在GTX 1650Ti上平均处理26张图像/s,在20 m范围内边界提取误差小于0.47 m。结果表明,本文提出的MSFA-Net能够显著降低导航误差,提高农机在野外作业中的感知稳定性。此外,该模型可以导出为ONNX或TensorFlow Lite格式,便于在嵌入式设备和现有农场导航系统上进行有效部署。
{"title":"Multi-scale feature alignment network for 19-class semantic segmentation in agricultural environments","authors":"Zhi-xin Yao ,&nbsp;Hao Wang ,&nbsp;Zhi-jun Meng ,&nbsp;Liang-liang Yang ,&nbsp;Tai-hong Zhang","doi":"10.1016/j.aiia.2025.11.008","DOIUrl":"10.1016/j.aiia.2025.11.008","url":null,"abstract":"<div><div>To improve environmental perception and ensure reliable agricultural machinery navigation during field transitions under unstructured farm road conditions, this study utilizes high-resolution RGB camera vision navigation technology to propose a Multi-Scale Feature Alignment Network (MSFA-Net) for 19-class semantic segmentation of agricultural environment, which includes information such as roads, pedestrians, and vehicles. MSFA-Net introduces two key innovations: the DASP module, which integrates multi-scale feature extraction with dual attention mechanisms (spatial and channel), and the MSFA architecture, which enables robust boundary extraction and mitigates interference from lighting variations and obstacles like vegetation. Compared to existing models, MSFA-Net uniquely combines efficient multi-scale feature extraction with real-time inference capabilities, achieving an mIoU of 84.46 % and an mPA of 96.10 %. For 512 × 512 input images, the model processes an average of 26 images/s on a GTX 1650Ti, with a boundary extraction error of less than 0.47 m within 20 m. These results indicate that the proposed MSFA-Net can significantly reduce navigation errors and improve the perception stability of agricultural machinery during field operations. Furthermore, the model can be exported to ONNX or TensorFlow Lite formats, facilitating efficient deployment on embedded devices and existing farm navigation systems.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"16 1","pages":"Pages 529-545"},"PeriodicalIF":12.4,"publicationDate":"2025-11-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145623317","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dual attention guided context-aware feature learning for residual unfilled grains detection on threshed rice panicles 双重注意引导的上下文感知特征学习在脱粒水稻穗部残留未填充粒检测中的应用
IF 12.4 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-11-15 DOI: 10.1016/j.aiia.2025.11.005
Yuhao Zhou , Xiao Feng , Shuqi Tang , Jinpeng Yang , Shaobin Chen , Xiangbao Meng , Zhanpeng Liang , Ruijun Ma , Long Qi
Accurate detection of residual unfilled grains on threshed rice panicles is a critical step in determining a reliable grain-setting rate, and holds significant potential for the development of high-quality rice strains. Recent deep learning-based techniques have been actively explored for discerning various types of objects. However, this detection task is challenging, as many objects are densely occluded by branches or other unfilled grains. Additionally, some unfilled grains are closely adjacent and exhibit small sizes, further complicating the detection process. To address these challenges, this paper proposes a novel Channel-global Spatial-local Dual Attention (CSDA) module, aimed at enhancing feature correlation learning and contextual information embedding. Specifically, the channel- and spatial-wise attention are deployed on two parallel branches, and incorporated with the global and local representation learning paradigm, respectively. Furthermore, we integrate the CSDA module with the backbone of an object detector, and refine the loss function and detection head using the Focaler-SIoU and tiny object prediction head. This enables the object detector to effectively differentiate residual unfilled grains from occlusions, and at the meantime, focusing on the subtle differences between closely adjacent and small-sized unfilled grains. Experimental results show that our work achieves superior detection performance versus other competitors with an [email protected] of 95.3 % (outperforming rivals by 1.5–32.6 %) and a frame rate of 154 FPS (outperforming rivals by 12–132 FPS), enjoying substantial potentials for practical applications.
准确检测脱粒后稻穗上的剩余未灌浆粒是确定可靠的结实率的关键步骤,对开发优质水稻品系具有重要的潜力。最近基于深度学习的技术已经被积极地用于识别各种类型的物体。然而,这种检测任务具有挑战性,因为许多物体被树枝或其他未填充的颗粒密集地遮挡。此外,一些未填充的颗粒紧密相邻且尺寸较小,进一步使检测过程复杂化。为了解决这些问题,本文提出了一种新的通道-全局空间-局部双重注意(CSDA)模块,旨在增强特征相关学习和上下文信息嵌入。具体来说,渠道和空间智慧的注意力被部署在两个平行的分支上,并分别与全局和局部表征学习范式相结合。此外,我们将CSDA模块与目标检测器的主干集成,并使用Focaler-SIoU和微小目标预测头来改进损失函数和检测头。这使得目标检测器能够有效地区分残差未填充颗粒和遮挡物,同时,能够专注于相邻颗粒和小尺寸未填充颗粒之间的细微差异。实验结果表明,我们的工作取得了优于其他竞争对手的检测性能,[email protected]的检测率为95.3%(优于竞争对手1.5 - 32.6%),帧率为154 FPS(优于竞争对手12-132 FPS),具有很大的实际应用潜力。
{"title":"Dual attention guided context-aware feature learning for residual unfilled grains detection on threshed rice panicles","authors":"Yuhao Zhou ,&nbsp;Xiao Feng ,&nbsp;Shuqi Tang ,&nbsp;Jinpeng Yang ,&nbsp;Shaobin Chen ,&nbsp;Xiangbao Meng ,&nbsp;Zhanpeng Liang ,&nbsp;Ruijun Ma ,&nbsp;Long Qi","doi":"10.1016/j.aiia.2025.11.005","DOIUrl":"10.1016/j.aiia.2025.11.005","url":null,"abstract":"<div><div>Accurate detection of residual unfilled grains on threshed rice panicles is a critical step in determining a reliable grain-setting rate, and holds significant potential for the development of high-quality rice strains. Recent deep learning-based techniques have been actively explored for discerning various types of objects. However, this detection task is challenging, as many objects are densely occluded by branches or other unfilled grains. Additionally, some unfilled grains are closely adjacent and exhibit small sizes, further complicating the detection process. To address these challenges, this paper proposes a novel Channel-global Spatial-local Dual Attention (CSDA) module, aimed at enhancing feature correlation learning and contextual information embedding. Specifically, the channel- and spatial-wise attention are deployed on two parallel branches, and incorporated with the global and local representation learning paradigm, respectively. Furthermore, we integrate the CSDA module with the backbone of an object detector, and refine the loss function and detection head using the Focaler-SIoU and tiny object prediction head. This enables the object detector to effectively differentiate residual unfilled grains from occlusions, and at the meantime, focusing on the subtle differences between closely adjacent and small-sized unfilled grains. Experimental results show that our work achieves superior detection performance versus other competitors with an [email protected] of 95.3 % (outperforming rivals by 1.5–32.6 %) and a frame rate of 154 FPS (outperforming rivals by 12–132 FPS), enjoying substantial potentials for practical applications.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"16 1","pages":"Pages 514-528"},"PeriodicalIF":12.4,"publicationDate":"2025-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145623316","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advancing lightweight and efficient detection of tomato main stems for edge device deployment 推进番茄主茎的轻量化和高效检测,用于边缘设备部署
IF 12.4 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-11-10 DOI: 10.1016/j.aiia.2025.10.016
Guohua Gao, Lifa Fang, Zihua Zhang, Jiahao Li
Automated pruning and defoliation of tomato plants are essential in modern cultivation systems for optimizing canopy structure, enhancing air circulation, and increasing yield. However, detecting main stems in the field faces significant challenges, like complex background interference, limited field of view, dense foliage occlusion, and curved stems. To address those challenges while ensuring hardware friendliness, computational efficiency, and real-time response, this study proposes a lightweight tomato main stem detection, optimisation, and deployment scheme. First, an efficient semi-automatic rotated bounding box annotation strategy is employed to segment the visible main stem segments, thus improving the adaptability to curved stems. Second, the lightweight network, YOLOR-Slim, is constructed to significantly reduce model complexity while maintaining detection performance through automated iterative pruning at the group-level of channel importance and a hybrid feature-based and logic-based knowledge distillation mechanism. Finally, an efficient and real-time main stem detection is achieved by deploying the model on inference engines and embedded platforms with various types and quantization bits. Experimental results showed that YOLOR-Slim achieved 87.5 % mAP@50, 1.9G Flops, 1.4 M parameters, and 7.4 ms inference time (pre-processing, inference, and post-processing) on the workstation, representing reductions of 2.8 %, 10.0 M, and 27.5G compared to the original model. After conversion with TensorRT, the inference time on Jetson Nano reached 57.6 ms, validating the operational efficiency and deployment applicability on edge devices. The YOLOR-Slim strikes a balance between inference speed, computational resources usage, and detection accuracy, providing a reliable perceptual foundation for automated pruning tasks in precision agriculture.
番茄植株的自动修剪和落叶在现代栽培系统中是优化树冠结构、促进空气流通和提高产量所必需的。然而,在野外检测主茎面临着复杂的背景干扰、有限的视野、茂密的树叶遮挡和弯曲的茎等重大挑战。为了应对这些挑战,同时确保硬件友好性、计算效率和实时响应,本研究提出了一种轻量级番茄主茎检测、优化和部署方案。首先,采用高效的半自动旋转包围框标注策略对可见的主茎段进行分割,提高了对弯曲茎的适应性;其次,构建轻量级网络yolo - slim,通过通道重要性组级的自动迭代剪枝和基于特征和基于逻辑的混合知识蒸馏机制,在保持检测性能的同时显著降低模型复杂性。最后,将该模型部署在推理引擎和具有各种类型和量化位的嵌入式平台上,实现了高效实时的主干检测。实验结果表明,yolo - slim在工作站上实现了87.5% mAP@50, 1.9G Flops, 1.4 M参数和7.4 ms推理时间(预处理,推理和后处理),与原始模型相比减少了2.8%,10.0 M和27.5G。经过TensorRT转换后,Jetson Nano上的推理时间达到57.6 ms,验证了在边缘设备上的运行效率和部署适用性。yolo - slim在推理速度、计算资源使用和检测精度之间取得了平衡,为精准农业中的自动修剪任务提供了可靠的感知基础。
{"title":"Advancing lightweight and efficient detection of tomato main stems for edge device deployment","authors":"Guohua Gao,&nbsp;Lifa Fang,&nbsp;Zihua Zhang,&nbsp;Jiahao Li","doi":"10.1016/j.aiia.2025.10.016","DOIUrl":"10.1016/j.aiia.2025.10.016","url":null,"abstract":"<div><div>Automated pruning and defoliation of tomato plants are essential in modern cultivation systems for optimizing canopy structure, enhancing air circulation, and increasing yield. However, detecting main stems in the field faces significant challenges, like complex background interference, limited field of view, dense foliage occlusion, and curved stems. To address those challenges while ensuring hardware friendliness, computational efficiency, and real-time response, this study proposes a lightweight tomato main stem detection, optimisation, and deployment scheme. First, an efficient semi-automatic rotated bounding box annotation strategy is employed to segment the visible main stem segments, thus improving the adaptability to curved stems. Second, the lightweight network, YOLOR-Slim, is constructed to significantly reduce model complexity while maintaining detection performance through automated iterative pruning at the group-level of channel importance and a hybrid feature-based and logic-based knowledge distillation mechanism. Finally, an efficient and real-time main stem detection is achieved by deploying the model on inference engines and embedded platforms with various types and quantization bits. Experimental results showed that YOLOR-Slim achieved 87.5 % mAP@50, 1.9G Flops, 1.4 M parameters, and 7.4 ms inference time (pre-processing, inference, and post-processing) on the workstation, representing reductions of 2.8 %, 10.0 M, and 27.5G compared to the original model. After conversion with TensorRT, the inference time on Jetson Nano reached 57.6 ms, validating the operational efficiency and deployment applicability on edge devices. The YOLOR-Slim strikes a balance between inference speed, computational resources usage, and detection accuracy, providing a reliable perceptual foundation for automated pruning tasks in precision agriculture.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"16 1","pages":"Pages 458-479"},"PeriodicalIF":12.4,"publicationDate":"2025-11-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145579034","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Energy-saving and stability-enhancing control for unmanned distributed drive electric plant protection vehicle based on active torque distribution 基于主动转矩分配的无人分布式驱动电动植保车节能增稳控制
IF 12.4 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-11-05 DOI: 10.1016/j.aiia.2025.11.004
Wenxiang Xu , Yejun Zhu , Maohua Xiao , Mengnan Liu , Liling Ye , Yanpeng Yang , Ze Liu
The distributed drive electric plant protection vehicle (DDEPPV), equipped with a unique four-wheel independent drive system, demonstrates excellent path-tracking capability and dynamic performance in agricultural environments. However, during actual field operations, issues such as severe tire slip, poor driving stability, high rollover risk, and excessive energy consumption often arise due to improper torque distribution. This study proposes an energy-efficient and stability-enhancing control method based on active torque distribution, aiming to improve both operational safety and system efficiency. A hierarchical control architecture is adopted: the upper-level controller employs a nonlinear model predictive control (NMPC) to achieve coordinated control of steering and yaw moment, enhancing lateral stability and ensuring operational safety. The lower-level controller implements a direct torque allocation method based on an adaptive-weight multi-objective twin delayed deep deterministic policy gradient (AW-MO-TD3) algorithm, enabling joint optimization of tire slip ratio and energy consumption. Real-vehicle tests were conducted under two typical field conditions, and the results show that compared with conventional methods, the proposed strategy significantly improves key performance metrics including tracking accuracy, vehicle stability, and energy efficiency. Specifically, stability was enhanced by 29.1 % and 41.4 %, while energy consumption was reduced by 19.8 % and 21.1 % under dry plowed terrain and muddy rice field conditions, respectively. This research provides technical support for the intelligent control of distributed drive electric agricultural vehicles.
分布式驱动电动植保车(DDEPPV)采用独特的四轮独立驱动系统,在农业环境中表现出优异的路径跟踪能力和动态性能。然而,在实际的现场作业中,由于扭矩分配不当,往往会出现严重的轮胎打滑、行驶稳定性差、侧翻风险高、能耗过大等问题。本文提出了一种基于主动转矩分配的节能增稳控制方法,旨在提高运行安全性和系统效率。采用层次控制体系结构,上层控制器采用非线性模型预测控制(NMPC),实现转向力矩与偏航力矩的协调控制,增强横向稳定性,保证运行安全。底层控制器实现了基于自适应权值多目标双延迟深度确定性策略梯度(AW-MO-TD3)算法的直接转矩分配方法,实现了轮胎打滑率和能耗的联合优化。在两种典型的现场条件下进行了实车试验,结果表明,与传统方法相比,该策略显著提高了跟踪精度、车辆稳定性和能源效率等关键性能指标。旱耕地形和泥泞稻田条件下,稳定性分别提高29.1%和41.4%,能耗分别降低19.8%和21.1%。本研究为分布式驱动农用电动车辆的智能控制提供了技术支持。
{"title":"Energy-saving and stability-enhancing control for unmanned distributed drive electric plant protection vehicle based on active torque distribution","authors":"Wenxiang Xu ,&nbsp;Yejun Zhu ,&nbsp;Maohua Xiao ,&nbsp;Mengnan Liu ,&nbsp;Liling Ye ,&nbsp;Yanpeng Yang ,&nbsp;Ze Liu","doi":"10.1016/j.aiia.2025.11.004","DOIUrl":"10.1016/j.aiia.2025.11.004","url":null,"abstract":"<div><div>The distributed drive electric plant protection vehicle (DDEPPV), equipped with a unique four-wheel independent drive system, demonstrates excellent path-tracking capability and dynamic performance in agricultural environments. However, during actual field operations, issues such as severe tire slip, poor driving stability, high rollover risk, and excessive energy consumption often arise due to improper torque distribution. This study proposes an energy-efficient and stability-enhancing control method based on active torque distribution, aiming to improve both operational safety and system efficiency. A hierarchical control architecture is adopted: the upper-level controller employs a nonlinear model predictive control (NMPC) to achieve coordinated control of steering and yaw moment, enhancing lateral stability and ensuring operational safety. The lower-level controller implements a direct torque allocation method based on an adaptive-weight multi-objective twin delayed deep deterministic policy gradient (AW-MO-TD3) algorithm, enabling joint optimization of tire slip ratio and energy consumption. Real-vehicle tests were conducted under two typical field conditions, and the results show that compared with conventional methods, the proposed strategy significantly improves key performance metrics including tracking accuracy, vehicle stability, and energy efficiency. Specifically, stability was enhanced by 29.1 % and 41.4 %, while energy consumption was reduced by 19.8 % and 21.1 % under dry plowed terrain and muddy rice field conditions, respectively. This research provides technical support for the intelligent control of distributed drive electric agricultural vehicles.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"16 1","pages":"Pages 495-513"},"PeriodicalIF":12.4,"publicationDate":"2025-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145623315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Two-year remote sensing and ground verification: Estimating chlorophyll content in winter wheat using UAV multi-spectral imagery 两年遥感与地面验证:利用无人机多光谱影像估算冬小麦叶绿素含量
IF 12.4 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-11-05 DOI: 10.1016/j.aiia.2025.10.017
Wenjie Ai , Guofeng Yang , Zhongren Li , Jiawei Du , Lingzhen Ye , Xuping Feng , Xiangping Jin , Yong He
Leaf chlorophyll content serves as a critical biophysical indicator for characterizing wheat growth status. Traditional measurement using a SPAD meter, while convenient, is hampered by its localized sampling, low efficiency, and destructive nature, making it unsuitable for high-throughput field applications. To overcome these constraints, this research developed a novel approach for assessing canopy SPAD values in winter wheat by leveraging multispectral imagery obtained from an unmanned aerial vehicle (UAV). The generalizability of this methodology was rigorously evaluated through a replication experiment conducted in a subsequent growing season. Throughout the study, canopy reflectance data were acquired across key phenological stages and paired with synchronized ground-based SPAD measurements to construct stage-specific estimation models. The acquired multispectral images were processed to remove soil background interference, from which 17 distinct vegetation indices and 8 texture features were subsequently extracted. An in-depth examination followed, aiming to clarify the evolving interplay of these features with SPAD values throughout growth phases. Among the vegetation indices, the Modified Climate Change Canopy Vegetation Index (MCCCI) displayed a “rise-and-decline” pattern across the season, aligning with the crop's intrinsic growth dynamics and establishing it as a robust and phonologically interpretable proxy. Texture features, particularly contrast and entropy, demonstrated notable associations with SPAD values, reaching their peak strength during the booting stage. Comparative evaluation of various predictive modeling techniques revealed that a Support Vector Regression (SVR) model integrating both vegetation indices and texture features yielded the highest estimation accuracy. This integrated model outperformed models based solely on spectral or textural data, improving estimation accuracy by 23.81 % and 22.48 %, respectively. The model's strong generalization capability was further confirmed on the independent validation dataset from the second year (RMSE = 2.54, R2 = 0.748). In summary, this study establishes an effective and transferable framework for non-destructively monitoring chlorophyll content in winter wheat canopies using UAV data.
叶片叶绿素含量是表征小麦生长状况的重要生物物理指标。使用SPAD仪表进行传统测量虽然方便,但由于其局部采样、低效率和破坏性,使其不适合高通量现场应用。为了克服这些限制,本研究开发了一种利用无人机(UAV)获得的多光谱图像来评估冬小麦冠层SPAD值的新方法。通过在随后的生长季节进行的重复实验,严格评估了该方法的普遍性。在整个研究过程中,获取了关键物候阶段的冠层反射率数据,并与同步的地面SPAD测量数据配对,构建了特定阶段的估算模型。对获取的多光谱图像进行去除土壤背景干扰的处理,提取出17种不同的植被指数和8种纹理特征。随后进行了深入的研究,旨在阐明这些特征在整个生长阶段与SPAD值之间不断变化的相互作用。在植被指数中,修正气候变化冠层植被指数(MCCCI)在整个季节中呈现出“上升-下降”的模式,与作物的内在生长动态一致,并建立了一个稳健的、可在音系上解释的指标。纹理特征,特别是对比度和熵,与SPAD值有显著的相关性,在启动阶段达到峰值。通过对各种预测建模技术的比较评估,发现结合植被指数和纹理特征的支持向量回归(SVR)模型的估计精度最高。该综合模型的估计精度分别提高了23.81%和22.48%,优于单纯基于光谱和纹理数据的模型。在第二年的独立验证数据集上进一步证实了模型较强的泛化能力(RMSE = 2.54, R2 = 0.748)。综上所述,本研究为利用无人机数据无损监测冬小麦冠层叶绿素含量建立了一个有效且可转移的框架。
{"title":"Two-year remote sensing and ground verification: Estimating chlorophyll content in winter wheat using UAV multi-spectral imagery","authors":"Wenjie Ai ,&nbsp;Guofeng Yang ,&nbsp;Zhongren Li ,&nbsp;Jiawei Du ,&nbsp;Lingzhen Ye ,&nbsp;Xuping Feng ,&nbsp;Xiangping Jin ,&nbsp;Yong He","doi":"10.1016/j.aiia.2025.10.017","DOIUrl":"10.1016/j.aiia.2025.10.017","url":null,"abstract":"<div><div>Leaf chlorophyll content serves as a critical biophysical indicator for characterizing wheat growth status. Traditional measurement using a SPAD meter, while convenient, is hampered by its localized sampling, low efficiency, and destructive nature, making it unsuitable for high-throughput field applications. To overcome these constraints, this research developed a novel approach for assessing canopy SPAD values in winter wheat by leveraging multispectral imagery obtained from an unmanned aerial vehicle (UAV). The generalizability of this methodology was rigorously evaluated through a replication experiment conducted in a subsequent growing season. Throughout the study, canopy reflectance data were acquired across key phenological stages and paired with synchronized ground-based SPAD measurements to construct stage-specific estimation models. The acquired multispectral images were processed to remove soil background interference, from which 17 distinct vegetation indices and 8 texture features were subsequently extracted. An in-depth examination followed, aiming to clarify the evolving interplay of these features with SPAD values throughout growth phases. Among the vegetation indices, the Modified Climate Change Canopy Vegetation Index (MCCCI) displayed a “rise-and-decline” pattern across the season, aligning with the crop's intrinsic growth dynamics and establishing it as a robust and phonologically interpretable proxy. Texture features, particularly contrast and entropy, demonstrated notable associations with SPAD values, reaching their peak strength during the booting stage. Comparative evaluation of various predictive modeling techniques revealed that a Support Vector Regression (SVR) model integrating both vegetation indices and texture features yielded the highest estimation accuracy. This integrated model outperformed models based solely on spectral or textural data, improving estimation accuracy by 23.81 % and 22.48 %, respectively. The model's strong generalization capability was further confirmed on the independent validation dataset from the second year (RMSE = 2.54, R<sup>2</sup> = 0.748). In summary, this study establishes an effective and transferable framework for non-destructively monitoring chlorophyll content in winter wheat canopies using UAV data.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"16 1","pages":"Pages 480-494"},"PeriodicalIF":12.4,"publicationDate":"2025-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145579035","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Artificial Intelligence in Agriculture
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1