首页 > 最新文献

Artificial Intelligence in Agriculture最新文献

英文 中文
FGPointKAN++ point cloud segmentation and adaptive key cutting plane recognition for cow body size measurement fgpointkan++点云分割和自适应关键切割平面识别的奶牛体型测量
IF 8.2 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-01 Epub Date: 2025-06-18 DOI: 10.1016/j.aiia.2025.06.003
Guoyuan Zhou , Wenhao Ye , Sheng Li , Jian Zhao , Zhiwen Wang , Guoliang Li , Jiawei Li
Accurate and efficient body size measurement is essential for health assessment and production management in modern animal husbandry. In order to realize the segmentation of the point clouds at the pixel-level and the accurate calculation of body size for the dairy cows in different postures, a segmentation model (FGPointKAN++) and an adaptive key cutting plane recognition (AKCPR) model are developed. FGPointKAN++ introduces FGE module and KAN that enhance local feature extraction and geometric consistency, significantly improving dairy cow part segmentation accuracy. The AKCPR utilizes adaptive plane fitting and dynamic orientation calibration to optimize the key body size measurement. The dairy cow body size parameters are then calculated based on the plane geometry features. The experimental results show that mIoU scores of 82.92 % and 83.24 % for the dairy cow pixel-level point cloud segmentation results. The calculated Mean Absolute Percentage Errors (MAPE) of Wither Height (WH), Body Width (BW), Chest Circumference (CC) and Abdominal Circumference (AC) are 2.07 %, 3.56 %, 2.24 % and 1.42 %, respectively. This method enables precise segmentation and automatic body size measurement of dairy cows in various walking postures, showing considerable potential for practical applications. It provides technical support for unmanned, intelligent, and precision farming, thereby enhancing animal welfare and improving economic efficiency.
准确、高效的体尺测量是现代畜牧业健康评价和生产管理的基础。为了实现点云的像素级分割和奶牛不同姿态体型的精确计算,开发了fgpointkan++分割模型和自适应关键切割平面识别(AKCPR)模型。fgpointkan++引入FGE模块和KAN,增强局部特征提取和几何一致性,显著提高奶牛部位分割精度。AKCPR利用自适应平面拟合和动态方向校准来优化关键体尺寸测量。然后根据平面几何特征计算奶牛体型参数。实验结果表明,奶牛像素级点云分割的mIoU分数分别为82.92%和83.24%。臀高(WH)、体宽(BW)、胸围(CC)和腹围(AC)的平均绝对百分比误差(MAPE)分别为2.07%、3.56%、2.24%和1.42%。该方法可实现奶牛不同行走姿势的精确分割和体型自动测量,具有较大的实际应用潜力。为无人化、智能化、精准化农业提供技术支持,从而提高动物福利,提高经济效益。
{"title":"FGPointKAN++ point cloud segmentation and adaptive key cutting plane recognition for cow body size measurement","authors":"Guoyuan Zhou ,&nbsp;Wenhao Ye ,&nbsp;Sheng Li ,&nbsp;Jian Zhao ,&nbsp;Zhiwen Wang ,&nbsp;Guoliang Li ,&nbsp;Jiawei Li","doi":"10.1016/j.aiia.2025.06.003","DOIUrl":"10.1016/j.aiia.2025.06.003","url":null,"abstract":"<div><div>Accurate and efficient body size measurement is essential for health assessment and production management in modern animal husbandry. In order to realize the segmentation of the point clouds at the pixel-level and the accurate calculation of body size for the dairy cows in different postures, a segmentation model (FGPointKAN++) and an adaptive key cutting plane recognition (AKCPR) model are developed. FGPointKAN++ introduces FGE module and KAN that enhance local feature extraction and geometric consistency, significantly improving dairy cow part segmentation accuracy. The AKCPR utilizes adaptive plane fitting and dynamic orientation calibration to optimize the key body size measurement. The dairy cow body size parameters are then calculated based on the plane geometry features. The experimental results show that mIoU scores of 82.92 % and 83.24 % for the dairy cow pixel-level point cloud segmentation results. The calculated Mean Absolute Percentage Errors (MAPE) of Wither Height (WH), Body Width (BW), Chest Circumference (CC) and Abdominal Circumference (AC) are 2.07 %, 3.56 %, 2.24 % and 1.42 %, respectively. This method enables precise segmentation and automatic body size measurement of dairy cows in various walking postures, showing considerable potential for practical applications. It provides technical support for unmanned, intelligent, and precision farming, thereby enhancing animal welfare and improving economic efficiency.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"15 4","pages":"Pages 783-801"},"PeriodicalIF":8.2,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144365220","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
VMGP: A unified variational auto-encoder based multi-task model for multi-phenotype, multi-environment, and cross-population genomic selection in plants VMGP:一个基于统一变分自编码器的多任务模型,用于植物的多表型、多环境和跨群体基因组选择
IF 8.2 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-01 Epub Date: 2025-06-24 DOI: 10.1016/j.aiia.2025.06.007
Xiangyu Zhao , Fuzhen Sun , Jinlong Li , Dongfeng Zhang , Qiusi Zhang , Zhongqiang Liu , Changwei Tan , Hongxiang Ma , Kaiyi Wang
Plant breeding stands as a cornerstone for agricultural productivity and the safeguarding of food security. The advent of Genomic Selection heralds a new epoch in breeding, characterized by its capacity to harness whole-genome variation for genomic prediction. This approach transcends the need for prior knowledge of genes associated with specific traits. Nonetheless, the vast dimensionality of genomic data juxtaposed with the relatively limited number of phenotypic samples often leads to the “curse of dimensionality”, where traditional statistical, machine learning, and deep learning methods are prone to overfitting and suboptimal predictive performance. To surmount this challenge, we introduce a unified Variational auto-encoder based Multi-task Genomic Prediction model (VMGP) that integrates self-supervised genomic compression and reconstruction with multiple prediction tasks. This approach provides a robust solution, offering a formidable predictive framework that has been rigorously validated across public datasets for wheat, rice, and maize. Our model demonstrates exceptional capabilities in multi-phenotype and multi-environment genomic prediction, successfully navigating the complexities of cross-population genomic selection and underscoring its unique strengths and utility. Furthermore, by integrating VMGP with model interpretability, we can effectively triage relevant single nucleotide polymorphisms, thereby enhancing prediction performance and proposing potential cost-effective genotyping solutions. The VMGP framework, with its simplicity, stable predictive prowess, and open-source code, is exceptionally well-suited for broad dissemination within plant breeding programs. It is particularly advantageous for breeders who prioritize phenotype prediction yet may not possess extensive knowledge in deep learning or proficiency in parameter tuning.
植物育种是农业生产力和保障粮食安全的基石。基因组选择的出现预示着育种的新时代,其特点是能够利用全基因组变异进行基因组预测。这种方法超越了对与特定性状相关的基因的先验知识的需要。尽管如此,庞大的基因组数据维度与相对有限的表型样本数量并置于一起,往往导致“维度诅咒”,传统的统计、机器学习和深度学习方法容易出现过拟合和次优预测性能。为了克服这一挑战,我们引入了一个统一的基于变分自编码器的多任务基因组预测模型(VMGP),该模型将自监督基因组压缩和重建与多个预测任务集成在一起。这种方法提供了一个强大的解决方案,提供了一个强大的预测框架,该框架已在小麦、水稻和玉米的公共数据集中得到严格验证。我们的模型展示了在多表型和多环境基因组预测方面的卓越能力,成功地驾驭了跨种群基因组选择的复杂性,并强调了其独特的优势和实用性。此外,通过将VMGP与模型可解释性相结合,我们可以有效地分类相关的单核苷酸多态性,从而提高预测性能并提出潜在的经济有效的基因分型解决方案。VMGP框架具有简单、稳定的预测能力和开源代码,非常适合在植物育种项目中广泛传播。这是特别有利的育种者优先考虑表型预测,但可能不具备广泛的知识,在深度学习或熟练掌握参数调整。
{"title":"VMGP: A unified variational auto-encoder based multi-task model for multi-phenotype, multi-environment, and cross-population genomic selection in plants","authors":"Xiangyu Zhao ,&nbsp;Fuzhen Sun ,&nbsp;Jinlong Li ,&nbsp;Dongfeng Zhang ,&nbsp;Qiusi Zhang ,&nbsp;Zhongqiang Liu ,&nbsp;Changwei Tan ,&nbsp;Hongxiang Ma ,&nbsp;Kaiyi Wang","doi":"10.1016/j.aiia.2025.06.007","DOIUrl":"10.1016/j.aiia.2025.06.007","url":null,"abstract":"<div><div>Plant breeding stands as a cornerstone for agricultural productivity and the safeguarding of food security. The advent of Genomic Selection heralds a new epoch in breeding, characterized by its capacity to harness whole-genome variation for genomic prediction. This approach transcends the need for prior knowledge of genes associated with specific traits. Nonetheless, the vast dimensionality of genomic data juxtaposed with the relatively limited number of phenotypic samples often leads to the “curse of dimensionality”, where traditional statistical, machine learning, and deep learning methods are prone to overfitting and suboptimal predictive performance. To surmount this challenge, we introduce a unified Variational auto-encoder based Multi-task Genomic Prediction model (VMGP) that integrates self-supervised genomic compression and reconstruction with multiple prediction tasks. This approach provides a robust solution, offering a formidable predictive framework that has been rigorously validated across public datasets for wheat, rice, and maize. Our model demonstrates exceptional capabilities in multi-phenotype and multi-environment genomic prediction, successfully navigating the complexities of cross-population genomic selection and underscoring its unique strengths and utility. Furthermore, by integrating VMGP with model interpretability, we can effectively triage relevant single nucleotide polymorphisms, thereby enhancing prediction performance and proposing potential cost-effective genotyping solutions. The VMGP framework, with its simplicity, stable predictive prowess, and open-source code, is exceptionally well-suited for broad dissemination within plant breeding programs. It is particularly advantageous for breeders who prioritize phenotype prediction yet may not possess extensive knowledge in deep learning or proficiency in parameter tuning.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"15 4","pages":"Pages 829-842"},"PeriodicalIF":8.2,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144534365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving accuracy and generalization in single kernel oil characteristics prediction in maize using NIR-HSI and a knowledge-injected spectral tabtransformer 利用NIR-HSI和知识注入谱表转换器提高玉米单粒油特性预测的准确性和泛化性
IF 8.2 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-01 Epub Date: 2025-06-11 DOI: 10.1016/j.aiia.2025.05.007
Anran Song , Xinyu Guo , Weiliang Wen , Chuanyu Wang , Shenghao Gu , Xiaoqian Chen , Juan Wang , Chunjiang Zhao
Near-infrared spectroscopy hyperspectral imaging (NIR-HSI) is widely used for seed component prediction due to its non-destructive and rapid nature. However, existing models often suffer from limited generalization, particularly when trained on small datasets, and there is a lack of effective deep learning (DL) models for spectral data analysis. To address these challenges, we propose the Knowledge-Injected Spectral TabTransformer (KIT-Spectral TabTransformer), an innovative adaptation of the traditional TabTransformer specifically designed for maize seeds. By integrating domain-specific knowledge, this approach enhances model training efficiency and predictive accuracy while reducing reliance on large datasets. The generalization capability of the model was rigorously validated through ten-fold cross-validation (10-CV). Compared to traditional machine learning methods, the attention-based CNN (ACNNR), and the Oil Characteristics Predictor Transformer (OCP-Transformer), the KIT-Spectral TabTransformer demonstrated superior performance in oil mass prediction, achieving Rp2= 0.9238 ± 0.0346, RMSEp = 0.1746 ± 0.0401. For oil content, Rp2= 0.9602 ± 0.0180 and RMSEp = 0.5301 ± 0.1446 on a dataset with oil content ranging from 0.81 % to 13.07 %. On the independent validation set, our model achieved R2 values of 0.7820 and 0.7586, along with RPD values of 2.1420 and 2.0355 in the two tasks, highlighting its strong prediction capability and potential for real-world application. These findings offer a potential method and direction for single seed oil prediction and related crop component analysis.
近红外光谱高光谱成像(NIR-HSI)因其无损、快速等优点被广泛应用于种子成分预测。然而,现有模型通常泛化有限,特别是在小数据集上训练时,并且缺乏用于光谱数据分析的有效深度学习(DL)模型。为了解决这些挑战,我们提出了知识注入光谱TabTransformer (KIT-Spectral TabTransformer),这是一种专门为玉米种子设计的传统TabTransformer的创新改编。通过集成领域特定知识,该方法提高了模型训练效率和预测准确性,同时减少了对大型数据集的依赖。通过10倍交叉验证(10-CV)严格验证了模型的泛化能力。与传统的机器学习方法、基于注意力的CNN (attention-based CNN, ACNNR)和油特性预测变压器(Oil characteristic Predictor Transformer, OCP-Transformer)相比,KIT-Spectral TabTransformer在油质量预测方面表现出更优异的性能,Rp2= 0.9238±0.0346,RMSEp = 0.1746±0.0401。在含油量为0.81% ~ 13.07%的数据集上,Rp2= 0.9602±0.0180,RMSEp = 0.5301±0.1446。在独立验证集上,我们的模型在两个任务中的R2值分别为0.7820和0.7586,RPD值分别为2.1420和2.0355,显示出了较强的预测能力和实际应用潜力。这些发现为单粒种子油脂预测及相关作物成分分析提供了潜在的方法和方向。
{"title":"Improving accuracy and generalization in single kernel oil characteristics prediction in maize using NIR-HSI and a knowledge-injected spectral tabtransformer","authors":"Anran Song ,&nbsp;Xinyu Guo ,&nbsp;Weiliang Wen ,&nbsp;Chuanyu Wang ,&nbsp;Shenghao Gu ,&nbsp;Xiaoqian Chen ,&nbsp;Juan Wang ,&nbsp;Chunjiang Zhao","doi":"10.1016/j.aiia.2025.05.007","DOIUrl":"10.1016/j.aiia.2025.05.007","url":null,"abstract":"<div><div>Near-infrared spectroscopy hyperspectral imaging (NIR-HSI) is widely used for seed component prediction due to its non-destructive and rapid nature. However, existing models often suffer from limited generalization, particularly when trained on small datasets, and there is a lack of effective deep learning (DL) models for spectral data analysis. To address these challenges, we propose the Knowledge-Injected Spectral TabTransformer (KIT-Spectral TabTransformer), an innovative adaptation of the traditional TabTransformer specifically designed for maize seeds. By integrating domain-specific knowledge, this approach enhances model training efficiency and predictive accuracy while reducing reliance on large datasets. The generalization capability of the model was rigorously validated through ten-fold cross-validation (10-CV). Compared to traditional machine learning methods, the attention-based CNN (ACNNR), and the Oil Characteristics Predictor Transformer (OCP-Transformer), the KIT-Spectral TabTransformer demonstrated superior performance in oil mass prediction, achieving <span><math><msubsup><mi>R</mi><mi>p</mi><mn>2</mn></msubsup></math></span>= 0.9238 ± 0.0346, RMSE<sub>p</sub> = 0.1746 ± 0.0401. For oil content, <span><math><msubsup><mi>R</mi><mi>p</mi><mn>2</mn></msubsup></math></span>= 0.9602 ± 0.0180 and RMSE<sub>p</sub> = 0.5301 ± 0.1446 on a dataset with oil content ranging from 0.81 % to 13.07 %. On the independent validation set, our model achieved <span><math><msup><mi>R</mi><mn>2</mn></msup></math></span> values of 0.7820 and 0.7586, along with RPD values of 2.1420 and 2.0355 in the two tasks, highlighting its strong prediction capability and potential for real-world application. These findings offer a potential method and direction for single seed oil prediction and related crop component analysis.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"15 4","pages":"Pages 802-815"},"PeriodicalIF":8.2,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144470916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fast extraction of navigation line and crop position based on LiDAR for cabbage crops 基于激光雷达的白菜导航线和作物位置快速提取
IF 8.2 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-01 Epub Date: 2025-04-04 DOI: 10.1016/j.aiia.2025.03.007
Jiang Pin , Tingfeng Guo , Minzi Xv , Xiangjun Zou , Wenwu Hu
This paper describes the design, algorithm development, and experimental verification of a precise spray perception system based on LiDAR were presented to address the issue that the navigation line extraction accuracy of self-propelled sprayers during field operations is low, resulting in wheels rolling over the ridges and excessive pesticide waste. A data processing framework was established for the precision spray perception system. Through data preprocessing, adaptive segmentation of crops and ditches, extraction of navigation lines and crop positioning, which were derived from the original LiDAR point cloud species. Data collection and analysis of the field environment of cabbages in different growth cycles were conducted to verify the stability of the precision spraying system. A controllable constant-speed experimental setup was established to compare the performance of LiDAR and depth camera in the same field environment. The experimental results show that at the self-propelled sprayer of speeds of 0.5 and 1 ms−1, the maximum lateral error is 0.112 m in a cabbage ridge environment with inter-row weeds, with an mean absolute lateral error of 0.059 m. The processing speed per frame does not exceed 43 ms. Compared to the machine vision algorithm, this method reduces the average processing time by 122 ms. The proposed system demonstrates superior accuracy, processing time, and robustness in crop identification and navigation line extraction compared to the machine vision system.
针对野外作业中自行式喷雾器导航线提取精度低、车轮滚过山脊、农药浪费过多等问题,介绍了基于激光雷达的精准喷雾感知系统的设计、算法开发和实验验证。建立了高精度喷雾感知系统的数据处理框架。通过数据预处理,从原始LiDAR点云物种中提取作物和沟渠的自适应分割、导航线提取和作物定位。通过对不同生长周期卷心菜田间环境的数据采集和分析,验证了精准喷洒系统的稳定性。为了比较激光雷达和深度相机在相同野外环境下的性能,建立了可控等速实验装置。实验结果表明,在速度为0.5和1 ms−1的自行式喷雾器中,在有行间杂草的白菜垄环境中,最大侧向误差为0.112 m,平均绝对侧向误差为0.059 m。每帧的处理速度不超过43毫秒。与机器视觉算法相比,该方法平均处理时间缩短了122 ms。与机器视觉系统相比,该系统在作物识别和导航线提取方面具有更高的精度、处理时间和鲁棒性。
{"title":"Fast extraction of navigation line and crop position based on LiDAR for cabbage crops","authors":"Jiang Pin ,&nbsp;Tingfeng Guo ,&nbsp;Minzi Xv ,&nbsp;Xiangjun Zou ,&nbsp;Wenwu Hu","doi":"10.1016/j.aiia.2025.03.007","DOIUrl":"10.1016/j.aiia.2025.03.007","url":null,"abstract":"<div><div>This paper describes the design, algorithm development, and experimental verification of a precise spray perception system based on LiDAR were presented to address the issue that the navigation line extraction accuracy of self-propelled sprayers during field operations is low, resulting in wheels rolling over the ridges and excessive pesticide waste. A data processing framework was established for the precision spray perception system. Through data preprocessing, adaptive segmentation of crops and ditches, extraction of navigation lines and crop positioning, which were derived from the original LiDAR point cloud species. Data collection and analysis of the field environment of cabbages in different growth cycles were conducted to verify the stability of the precision spraying system. A controllable constant-speed experimental setup was established to compare the performance of LiDAR and depth camera in the same field environment. The experimental results show that at the self-propelled sprayer of speeds of 0.5 and 1 ms<sup>−1</sup>, the maximum lateral error is 0.112 m in a cabbage ridge environment with inter-row weeds, with an mean absolute lateral error of 0.059 m. The processing speed per frame does not exceed 43 ms. Compared to the machine vision algorithm, this method reduces the average processing time by 122 ms. The proposed system demonstrates superior accuracy, processing time, and robustness in crop identification and navigation line extraction compared to the machine vision system.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"15 4","pages":"Pages 686-695"},"PeriodicalIF":8.2,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144254490","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MSNet: A multispectral-image driven rapeseed canopy instance segmentation network 一个多光谱图像驱动的油菜籽冠层实例分割网络
IF 8.2 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-01 Epub Date: 2025-05-31 DOI: 10.1016/j.aiia.2025.05.008
Yuang Yang, Xiaole Wang, Fugui Zhang, Zhenchao Wu, Yu Wang, Yujie Liu, Xuan Lv, Bowen Luo, Liqing Chen, Yang Yang
Precise detection of rapeseed and the growth of its canopy area are crucial phenotypic indicators of its growth status. Achieving accurate identification of the rapeseed target and its growth region provides significant data support for phenotypic analysis and breeding research. However, in natural field environments, rapeseed detection remains a substantial challenge due to the limited feature representation capabilities of RGB-only modalities. To address this challenge, this study proposes a dual-modal instance segmentation network, MSNet, based on YOLOv11n-seg, integrating both RGB and Near-Infrared (NIR) modalities. The main improvements of this network include three different fusion location strategies (frontend fusion, mid-stage fusion, and backend fusion) and the newly introduced Hierarchical Attention Fusion Block (HAFB) for multimodal feature fusion. Comparative experiments on fusion locations indicate that the mid-stage fusion strategy achieves the best balance between detection accuracy and parameter efficiency. Compared to the baseline network, the mAP50:95 improvement can reach up to 3.5 %. After introducing the HAFB module, the MSNet-H-HAFB model demonstrates a 6.5 % increase in mAP50:95 relative to the baseline network, with less than a 38 % increase in parameter count. It is noteworthy that the mid-stage fusion consistently delivered the best detection performance in all experiments, providing clear design guidance for selecting fusion locations in future multimodal networks. In addition, comparisons with various RGB-only instance segmentation models show that all the proposed MSNet-HAFB fusion models significantly outperform single-modal models in rapeseed count detection tasks, confirming the potential advantages of multispectral fusion strategies in agricultural target recognition. Finally, the MSNet was applied in an agricultural case study, including vegetation index level analysis and frost damage classification. The results show that ZN6–2836 and ZS11 were predicted as potential superior varieties, and the EVI2 vegetation index achieved the best performance in rapeseed frost damage classification.
油菜籽的生长状况及其冠层面积的精确检测是反映油菜籽生长状况的重要表型指标。实现油菜靶点及其生长区域的准确鉴定,为表型分析和育种研究提供了重要的数据支持。然而,在自然野外环境中,由于仅rgb模式的特征表示能力有限,油菜籽检测仍然是一个重大挑战。为了解决这一挑战,本研究提出了一种基于YOLOv11n-seg的双模态实例分割网络MSNet,该网络集成了RGB和近红外(NIR)模式。该网络的主要改进包括三种不同的融合定位策略(前端融合、中期融合和后端融合)和新引入的用于多模态特征融合的分层注意融合块(HAFB)。融合位置的对比实验表明,中期融合策略在检测精度和参数效率之间达到了最佳平衡。与基线网络相比,mAP50:95的改进可达3.5%。在引入HAFB模块后,MSNet-H-HAFB模型显示,相对于基线网络,mAP50:95增加了6.5%,参数数量增加了不到38%。值得注意的是,中期融合在所有实验中始终提供了最佳的检测性能,为未来多模态网络中融合位置的选择提供了明确的设计指导。此外,与各种仅rgb实例分割模型的比较表明,所提出的MSNet-HAFB融合模型在油菜籽计数检测任务中都明显优于单模态模型,证实了多光谱融合策略在农业目标识别中的潜在优势。最后,将MSNet应用于农业案例研究,包括植被指数水平分析和霜冻灾害分类。结果表明,ZN6-2836和ZS11被预测为潜在优势品种,EVI2植被指数在油菜籽冻害分类中表现最佳。
{"title":"MSNet: A multispectral-image driven rapeseed canopy instance segmentation network","authors":"Yuang Yang,&nbsp;Xiaole Wang,&nbsp;Fugui Zhang,&nbsp;Zhenchao Wu,&nbsp;Yu Wang,&nbsp;Yujie Liu,&nbsp;Xuan Lv,&nbsp;Bowen Luo,&nbsp;Liqing Chen,&nbsp;Yang Yang","doi":"10.1016/j.aiia.2025.05.008","DOIUrl":"10.1016/j.aiia.2025.05.008","url":null,"abstract":"<div><div>Precise detection of rapeseed and the growth of its canopy area are crucial phenotypic indicators of its growth status. Achieving accurate identification of the rapeseed target and its growth region provides significant data support for phenotypic analysis and breeding research. However, in natural field environments, rapeseed detection remains a substantial challenge due to the limited feature representation capabilities of RGB-only modalities. To address this challenge, this study proposes a dual-modal instance segmentation network, MSNet, based on YOLOv11n-seg, integrating both RGB and Near-Infrared (NIR) modalities. The main improvements of this network include three different fusion location strategies (frontend fusion, mid-stage fusion, and backend fusion) and the newly introduced Hierarchical Attention Fusion Block (HAFB) for multimodal feature fusion. Comparative experiments on fusion locations indicate that the mid-stage fusion strategy achieves the best balance between detection accuracy and parameter efficiency. Compared to the baseline network, the <em>mAP50:95</em> improvement can reach up to 3.5 %. After introducing the HAFB module, the MSNet-H-HAFB model demonstrates a 6.5 % increase in <em>mAP50:95</em> relative to the baseline network, with less than a 38 % increase in parameter count. It is noteworthy that the mid-stage fusion consistently delivered the best detection performance in all experiments, providing clear design guidance for selecting fusion locations in future multimodal networks. In addition, comparisons with various RGB-only instance segmentation models show that all the proposed MSNet-HAFB fusion models significantly outperform single-modal models in rapeseed count detection tasks, confirming the potential advantages of multispectral fusion strategies in agricultural target recognition. Finally, the MSNet was applied in an agricultural case study, including vegetation index level analysis and frost damage classification. The results show that ZN6–2836 and ZS11 were predicted as potential superior varieties, and the EVI2 vegetation index achieved the best performance in rapeseed frost damage classification.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"15 4","pages":"Pages 642-658"},"PeriodicalIF":8.2,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144231062","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An autonomous navigation method for field phenotyping robot based on ground-air collaboration 基于地空协同的现场分型机器人自主导航方法
IF 8.2 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-01 Epub Date: 2025-05-30 DOI: 10.1016/j.aiia.2025.05.005
Zikang Zhang , Zhengda Li , Meng Yang , Jiale Cui , Yang Shao , Youchun Ding , Wanneng Yang , Wen Qiao , Peng Song
High-throughput phenotyping collection technology is important in affecting the efficiency of crop breeding. This study introduces a novel autonomous navigation method for phenotyping robots that leverages ground-air collaboration to meet the demands of unmanned crop phenotypic data collection. The proposed method employs a UAV equipped with a Real-Time Kinematic (RTK) module for the construction of high-precision Field maps. It utilizes SegFormor-B0 semantic segmentation models to detect crop rows, and extracts key coordinate points of these rows, and generates navigation paths for the phenotyping robots by mapping these points to actual geographic coordinates. Furthermore, an adaptive controller based on the Pure Pursuit algorithm is proposed, which dynamically adjusts the steering angle of the phenotyping robot in real-time, according to the distance (d), angular deviation (α) and the lateral deviation (ey) between the robot's current position and its target position. This enables the robot to accurately trace paths in field environments. The results demonstrate that the mean absolute error (MAE) of the proposed method in extracting the centerline of potted plants area's rows is 2.83 cm, and the cropland's rows is 4.51 cm. The majority of global path tracking errors stay within 2 cm. In the potted plants area, 99.1 % of errors lie within this range, with a mean absolute error of 0.62 cm and a maximum error of 2.59 cm. In the cropland, 72.4 % of errors remain within this range, with a mean absolute error of 1.51 cm and a maximum error of 4.22 cm. Compared with traditional GNSS-based navigation methods and single vision methods, this method shows significant advantages in adapting to the dynamic growth of crops and complex field environments, which not only ensures that the phenotyping robot accurately travels along the crop rows during field operations to avoid damage to the crops, but also provides an efficient and accurate means of data acquisition for crop phenotyping.
高通量表型收集技术是影响作物育种效率的重要技术之一。本研究介绍了一种新型的表型机器人自主导航方法,该方法利用地空协作来满足无人作物表型数据收集的需求。该方法采用一种配备实时运动学(RTK)模块的无人机来构建高精度的野外地图。它利用SegFormor-B0语义分割模型检测作物行,提取这些行的关键坐标点,并将这些点映射到实际地理坐标,为表型机器人生成导航路径。在此基础上,提出了一种基于Pure Pursuit算法的自适应控制器,根据机器人当前位置与目标位置之间的距离(d)、角度偏差(α)和横向偏差(ey),实时动态调整表型机器人的转向角度。这使机器人能够在现场环境中准确地跟踪路径。结果表明,该方法提取盆栽区行中心线的平均绝对误差(MAE)为2.83 cm,农田行中心线的平均绝对误差为4.51 cm。大多数全局路径跟踪误差保持在2cm以内。在盆栽区域,99.1%的误差在此范围内,平均绝对误差为0.62 cm,最大误差为2.59 cm。在农田中,72.4%的误差保持在该范围内,平均绝对误差为1.51 cm,最大误差为4.22 cm。与传统的基于gnss的导航方法和单视觉方法相比,该方法在适应作物的动态生长和复杂的田间环境方面具有明显的优势,不仅保证了表型机器人在田间作业中准确地沿着作物行移动,避免对作物造成损害,而且为作物表型分析提供了一种高效、准确的数据采集手段。
{"title":"An autonomous navigation method for field phenotyping robot based on ground-air collaboration","authors":"Zikang Zhang ,&nbsp;Zhengda Li ,&nbsp;Meng Yang ,&nbsp;Jiale Cui ,&nbsp;Yang Shao ,&nbsp;Youchun Ding ,&nbsp;Wanneng Yang ,&nbsp;Wen Qiao ,&nbsp;Peng Song","doi":"10.1016/j.aiia.2025.05.005","DOIUrl":"10.1016/j.aiia.2025.05.005","url":null,"abstract":"<div><div>High-throughput phenotyping collection technology is important in affecting the efficiency of crop breeding. This study introduces a novel autonomous navigation method for phenotyping robots that leverages ground-air collaboration to meet the demands of unmanned crop phenotypic data collection. The proposed method employs a UAV equipped with a Real-Time Kinematic (RTK) module for the construction of high-precision Field maps. It utilizes SegFormor-B0 semantic segmentation models to detect crop rows, and extracts key coordinate points of these rows, and generates navigation paths for the phenotyping robots by mapping these points to actual geographic coordinates. Furthermore, an adaptive controller based on the Pure Pursuit algorithm is proposed, which dynamically adjusts the steering angle of the phenotyping robot in real-time, according to the distance (<span><math><mi>d</mi></math></span>), angular deviation (<span><math><mi>α</mi></math></span>) and the lateral deviation (<span><math><msub><mi>e</mi><mi>y</mi></msub></math></span>) between the robot's current position and its target position. This enables the robot to accurately trace paths in field environments. The results demonstrate that the mean absolute error (MAE) of the proposed method in extracting the centerline of potted plants area's rows is 2.83 cm, and the cropland's rows is 4.51 cm. The majority of global path tracking errors stay within 2 cm. In the potted plants area, 99.1 % of errors lie within this range, with a mean absolute error of 0.62 cm and a maximum error of 2.59 cm. In the cropland, 72.4 % of errors remain within this range, with a mean absolute error of 1.51 cm and a maximum error of 4.22 cm. Compared with traditional GNSS-based navigation methods and single vision methods, this method shows significant advantages in adapting to the dynamic growth of crops and complex field environments, which not only ensures that the phenotyping robot accurately travels along the crop rows during field operations to avoid damage to the crops, but also provides an efficient and accurate means of data acquisition for crop phenotyping.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"15 4","pages":"Pages 610-621"},"PeriodicalIF":8.2,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144194569","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A new tool to improve the computation of animal kinetic activity indices in precision poultry farming 一种改进精密家禽养殖动物动力活动指数计算的新工具
IF 8.2 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-01 Epub Date: 2025-03-22 DOI: 10.1016/j.aiia.2025.03.005
Alberto Carraro , Mattia Pravato , Francesco Marinello , Francesco Bordignon , Angela Trocino , Gerolamo Xiccato , Andrea Pezzuolo
Precision Livestock Farming (PLF) emerges as a promising solution for revolutionising farming by enabling real-time automated monitoring of animals through smart technologies. PLF provides farmers with precise data to enhance farm management, increasing productivity and profitability. For instance, it allows for non-intrusive health assessments, contributing to maintaining a healthy herd while reducing stress associated with handling. In the poultry sector, image analysis can be utilised to monitor and analyse the behaviour of each hen in real time. Researchers have recently used machine learning algorithms to monitor the behaviour, health, and positioning of hens through computer vision techniques. Convolutional neural networks, a type of deep learning algorithm, have been utilised for image analysis to identify and categorise various hen behaviours and track specific activities like feeding and drinking. This research presents an automated system for analysing laying hen movement using video footage from surveillance cameras. With a customised implementation of object tracking, the system can efficiently process hundreds of hours of videos while maintaining high measurement precision. Its modular implementation adapts well to optimally exploit the GPU computing capabilities of the hardware platform it is running on. The use of this system is beneficial for both real-time monitoring and post-processing, contributing to improved monitoring capabilities in precision livestock farming.
精准畜牧业(PLF)通过智能技术实现对动物的实时自动化监控,成为一种有前途的农业革命解决方案。PLF为农民提供精确的数据,以加强农场管理,提高生产力和盈利能力。例如,它允许进行非侵入性健康评估,有助于保持健康的牛群,同时减少与处理相关的压力。在家禽业,图像分析可用于实时监测和分析每只母鸡的行为。研究人员最近使用机器学习算法通过计算机视觉技术来监测母鸡的行为、健康和定位。卷积神经网络是一种深度学习算法,已被用于图像分析,以识别和分类母鸡的各种行为,并跟踪诸如喂食和饮水等特定活动。本研究提出了一种利用监控摄像机录像片段分析蛋鸡运动的自动化系统。通过定制的目标跟踪实现,该系统可以有效地处理数百小时的视频,同时保持高测量精度。它的模块化实现很好地适应了它所运行的硬件平台的GPU计算能力。该系统的使用有利于实时监测和后期处理,有助于提高精准畜牧业的监测能力。
{"title":"A new tool to improve the computation of animal kinetic activity indices in precision poultry farming","authors":"Alberto Carraro ,&nbsp;Mattia Pravato ,&nbsp;Francesco Marinello ,&nbsp;Francesco Bordignon ,&nbsp;Angela Trocino ,&nbsp;Gerolamo Xiccato ,&nbsp;Andrea Pezzuolo","doi":"10.1016/j.aiia.2025.03.005","DOIUrl":"10.1016/j.aiia.2025.03.005","url":null,"abstract":"<div><div>Precision Livestock Farming (PLF) emerges as a promising solution for revolutionising farming by enabling real-time automated monitoring of animals through smart technologies. PLF provides farmers with precise data to enhance farm management, increasing productivity and profitability. For instance, it allows for non-intrusive health assessments, contributing to maintaining a healthy herd while reducing stress associated with handling. In the poultry sector, image analysis can be utilised to monitor and analyse the behaviour of each hen in real time. Researchers have recently used machine learning algorithms to monitor the behaviour, health, and positioning of hens through computer vision techniques. Convolutional neural networks, a type of deep learning algorithm, have been utilised for image analysis to identify and categorise various hen behaviours and track specific activities like feeding and drinking. This research presents an automated system for analysing laying hen movement using video footage from surveillance cameras. With a customised implementation of object tracking, the system can efficiently process hundreds of hours of videos while maintaining high measurement precision. Its modular implementation adapts well to optimally exploit the GPU computing capabilities of the hardware platform it is running on. The use of this system is beneficial for both real-time monitoring and post-processing, contributing to improved monitoring capabilities in precision livestock farming.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"15 4","pages":"Pages 659-670"},"PeriodicalIF":8.2,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144253342","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Recognizing and localizing chicken behaviors in videos based on spatiotemporal feature learning 基于时空特征学习的视频鸡行为识别与定位
IF 8.2 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-01 Epub Date: 2025-06-21 DOI: 10.1016/j.aiia.2025.06.006
Yilei Hu , Jinyang Xu , Zhichao Gou , Di Cui
Timely acquisition of chicken behavioral information is crucial for assessing chicken health status and production performance. Video-based behavior recognition has emerged as a primary technique for obtaining such information due to its accuracy and robustness. Video-based models generally predict a single behavior from a single video segment of a fixed duration. However, during periods of high activity in poultry, behavior transition may occur within a video segment, and existing models often fail to capture such transitions effectively. This limitation highlights the insufficient temporal resolution of video-based behavior recognition models. This study presents a chicken behavior recognition and localization model, CBLFormer, which is based on spatiotemporal feature learning. The model was designed to recognize behaviors that occur before and after transitions in video segments and to localize the corresponding time interval for each behavior. An improved transformer block, the cascade encoder-decoder network (CEDNet), a transformer-based head, and weighted distance intersection over union (WDIoU) loss were integrated into CBLFormer to enhance the model's ability to distinguish between different behavior categories and locate behavior boundaries. For the training and testing of CBLFormer, a dataset was created by collecting videos from 320 chickens across different ages and rearing densities. The results showed that CBLFormer achieved a [email protected]:0.95 of 98.34 % on the test set. The integration of CEDNet contributed the most to the performance improvement of CBLFormer. The visualization results confirmed that the model effectively captured the behavioral boundaries of chickens and correctly recognized behavior categories. The transfer learning results demonstrated that the model is applicable to chicken behavior recognition and localization tasks in real-world poultry farms. The proposed method handles cases where poultry behavior transitions occur within the video segment and improves the temporal resolution of video-based behavior recognition models.
及时获取鸡的行为信息对评估鸡的健康状况和生产性能至关重要。基于视频的行为识别由于其准确性和鲁棒性而成为获取此类信息的主要技术。基于视频的模型通常从固定时长的单个视频片段中预测单个行为。然而,在家禽的高活动期间,行为转变可能发生在视频片段中,现有模型通常无法有效捕捉这种转变。这一限制突出了基于视频的行为识别模型的时间分辨率不足。提出了一种基于时空特征学习的鸡行为识别与定位模型CBLFormer。该模型旨在识别视频片段中过渡前后发生的行为,并为每个行为定位相应的时间间隔。将改进的变压器块、级联编码器-解码器网络(CEDNet)、基于变压器的磁头和加权距离交联(WDIoU)损失集成到CBLFormer中,以增强模型区分不同行为类别和定位行为边界的能力。为了训练和测试CBLFormer,通过收集320只鸡不同年龄和饲养密度的视频,创建了一个数据集。结果表明,CBLFormer在98.34%的测试集中实现了[email protected]:0.95。集成CEDNet对CBLFormer的性能提升贡献最大。可视化结果证实,该模型有效捕获了鸡的行为边界,正确识别了鸡的行为类别。迁移学习结果表明,该模型适用于现实家禽养殖场中鸡的行为识别和定位任务。该方法处理了家禽行为在视频片段内发生转变的情况,并提高了基于视频的行为识别模型的时间分辨率。
{"title":"Recognizing and localizing chicken behaviors in videos based on spatiotemporal feature learning","authors":"Yilei Hu ,&nbsp;Jinyang Xu ,&nbsp;Zhichao Gou ,&nbsp;Di Cui","doi":"10.1016/j.aiia.2025.06.006","DOIUrl":"10.1016/j.aiia.2025.06.006","url":null,"abstract":"<div><div>Timely acquisition of chicken behavioral information is crucial for assessing chicken health status and production performance. Video-based behavior recognition has emerged as a primary technique for obtaining such information due to its accuracy and robustness. Video-based models generally predict a single behavior from a single video segment of a fixed duration. However, during periods of high activity in poultry, behavior transition may occur within a video segment, and existing models often fail to capture such transitions effectively. This limitation highlights the insufficient temporal resolution of video-based behavior recognition models. This study presents a chicken behavior recognition and localization model, CBLFormer, which is based on spatiotemporal feature learning. The model was designed to recognize behaviors that occur before and after transitions in video segments and to localize the corresponding time interval for each behavior. An improved transformer block, the cascade encoder-decoder network (CEDNet), a transformer-based head, and weighted distance intersection over union (WDIoU) loss were integrated into CBLFormer to enhance the model's ability to distinguish between different behavior categories and locate behavior boundaries. For the training and testing of CBLFormer, a dataset was created by collecting videos from 320 chickens across different ages and rearing densities. The results showed that CBLFormer achieved a [email protected]:0.95 of 98.34 % on the test set. The integration of CEDNet contributed the most to the performance improvement of CBLFormer. The visualization results confirmed that the model effectively captured the behavioral boundaries of chickens and correctly recognized behavior categories. The transfer learning results demonstrated that the model is applicable to chicken behavior recognition and localization tasks in real-world poultry farms. The proposed method handles cases where poultry behavior transitions occur within the video segment and improves the temporal resolution of video-based behavior recognition models.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"15 4","pages":"Pages 816-828"},"PeriodicalIF":8.2,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144480996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Technical study on the efficiency and models of weed control methods using unmanned ground vehicles: A review 无人驾驶地面车辆除草效率与模型技术研究综述
IF 8.2 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-12-01 Epub Date: 2025-05-28 DOI: 10.1016/j.aiia.2025.05.003
Evans K. Wiafe, Kelvin Betitame, Billy G. Ram, Xin Sun
As precision agriculture evolves, unmanned ground vehicles (UGVs) have become an essential tool for improving weed management techniques, offering automated and targeted methods that obviously reduce the reliance on manual labor and blanket herbicide applications. Several papers on UGV-based weed control methods have been published in recent years, yet there is no explicit attempt to systematically study these papers to discuss these weed control methods, UGVs adopted, and their key components, and how they impact the environment and economy. Therefore, the objective of this study was to present a systematic review that involves the efficiency and types of weed control methods deployed in UGVs, including mechanical weeding, targeted herbicide application, thermal/flaming weeding, and laser weeding in the last 2 decades. For this purpose, a thorough literature review was conducted, analyzing 68 relevant articles on weed control methods for UGVs. The study found that the research focus on using UGVs in mechanical weeding has been more dominant, followed by target or precision spraying/ chemical weeding, with hybrid weeding systems quickly emerging. The effectiveness of UGVs for weed control is hinged on the accuracy of their navigation and weed detection technologies, which are influenced heavily by environmental conditions, including lighting, weather, uneven terrain, and weed and crop density. Also, there is a shift from using traditional machine learning (ML) algorithms to deep learning neural networks, including convolutional neural networks (CNNs) and recurrent neural networks (RNNs), for weed detection algorithm development due to their potential to work in complex environments. Finally, trials of most UGVs have limited documentation or lack extensive trials under various conditions, such as varying soil types, crop fields, topography, field geometry, and annual weather conditions. This review paper serves as an in-depth update on UGVs in weed management for farmers, researchers, robotic technology industry players, and AI enthusiasts, helping to further foster collaborative efforts to develop new ideas and advance this revolutionary technique in modern agriculture.
随着精准农业的发展,无人驾驶地面车辆(ugv)已成为改善杂草管理技术的重要工具,提供自动化和有针对性的方法,明显减少了对人工劳动和地域性除草剂应用的依赖。近年来,关于基于ugv的杂草控制方法的论文已经发表了一些,但没有明确的尝试系统地研究这些论文,讨论这些杂草控制方法,采用的ugv,及其关键组成部分,以及它们对环境和经济的影响。因此,本研究的目的是对过去20年来ugv中使用的杂草控制方法的效率和类型进行系统回顾,包括机械除草、靶向除草剂施用、热/火焰除草和激光除草。为此,我们进行了全面的文献综述,分析了68篇关于ugv杂草控制方法的相关文章。研究发现,在机械除草中使用ugv的研究已经占据主导地位,其次是目标或精确喷洒/化学除草,混合除草系统迅速出现。ugv的杂草控制效果取决于其导航和杂草检测技术的准确性,而这些技术受环境条件的影响很大,包括光照、天气、不平坦地形、杂草和作物密度。此外,由于杂草检测算法具有在复杂环境中工作的潜力,因此从使用传统机器学习(ML)算法转向使用深度学习神经网络,包括卷积神经网络(cnn)和循环神经网络(rnn)。最后,大多数ugv的试验文件有限,或者缺乏在各种条件下的广泛试验,例如不同的土壤类型、作物田地、地形、田地几何形状和年度天气条件。这篇综述论文为农民、研究人员、机器人技术行业参与者和人工智能爱好者提供了ugv在杂草管理方面的深入更新,有助于进一步促进合作,开发新思路,推进现代农业中这一革命性技术的发展。
{"title":"Technical study on the efficiency and models of weed control methods using unmanned ground vehicles: A review","authors":"Evans K. Wiafe,&nbsp;Kelvin Betitame,&nbsp;Billy G. Ram,&nbsp;Xin Sun","doi":"10.1016/j.aiia.2025.05.003","DOIUrl":"10.1016/j.aiia.2025.05.003","url":null,"abstract":"<div><div>As precision agriculture evolves, unmanned ground vehicles (UGVs) have become an essential tool for improving weed management techniques, offering automated and targeted methods that obviously reduce the reliance on manual labor and blanket herbicide applications. Several papers on UGV-based weed control methods have been published in recent years, yet there is no explicit attempt to systematically study these papers to discuss these weed control methods, UGVs adopted, and their key components, and how they impact the environment and economy. Therefore, the objective of this study was to present a systematic review that involves the efficiency and types of weed control methods deployed in UGVs, including mechanical weeding, targeted herbicide application, thermal/flaming weeding, and laser weeding in the last 2 decades. For this purpose, a thorough literature review was conducted, analyzing 68 relevant articles on weed control methods for UGVs. The study found that the research focus on using UGVs in mechanical weeding has been more dominant, followed by target or precision spraying/ chemical weeding, with hybrid weeding systems quickly emerging. The effectiveness of UGVs for weed control is hinged on the accuracy of their navigation and weed detection technologies, which are influenced heavily by environmental conditions, including lighting, weather, uneven terrain, and weed and crop density. Also, there is a shift from using traditional machine learning (ML) algorithms to deep learning neural networks, including convolutional neural networks (CNNs) and recurrent neural networks (RNNs), for weed detection algorithm development due to their potential to work in complex environments. Finally, trials of most UGVs have limited documentation or lack extensive trials under various conditions, such as varying soil types, crop fields, topography, field geometry, and annual weather conditions. This review paper serves as an in-depth update on UGVs in weed management for farmers, researchers, robotic technology industry players, and AI enthusiasts, helping to further foster collaborative efforts to develop new ideas and advance this revolutionary technique in modern agriculture.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"15 4","pages":"Pages 622-641"},"PeriodicalIF":8.2,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144203295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Transformer-based audio-visual multimodal fusion for fine-grained recognition of individual sow nursing behaviour 基于变压器的视听多模态融合技术,用于精细识别母猪的哺乳行为
IF 8.2 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2025-09-01 Epub Date: 2025-04-08 DOI: 10.1016/j.aiia.2025.03.006
Yuqing Yang , Chengguo Xu , Wenhao Hou , Alan G. McElligott , Kai Liu , Yueju Xue
Nursing behaviour and the calling-to-nurse sound are crucial indicators for assessing sow maternal behaviour and nursing status. However, accurately identifying these behaviours for individual sows in complex indoor pig housing is challenging due to factors such as variable lighting, rail obstructions, and interference from other sows' calls. Multimodal fusion, which integrates audio and visual data, has proven to be an effective approach for improving accuracy and robustness in complex scenarios. In this study, we designed an audio-visual data acquisition system that includes a camera for synchronised audio and video capture, along with a custom-developed sound source localisation system that leverages a sound sensor to track sound direction. Specifically, we proposed a novel transformer-based audio-visual multimodal fusion (TMF) framework for recognising fine-grained sow nursing behaviour with or without the calling-to-nurse sound. Initially, a unimodal self-attention enhancement (USE) module was employed to augment video and audio features with global contextual information. Subsequently, we developed an audio-visual interaction enhancement (AVIE) module to compress relevant information and reduce noise using the information bottleneck principle. Moreover, we presented an adaptive dynamic decision fusion strategy to optimise the model's performance by focusing on the most relevant features in each modality. Finally, we comprehensively identified fine-grained nursing behaviours by integrating audio and fused information, while incorporating angle information from the real-time sound source localisation system to accurately determine whether the sound cues originate from the target sow. Our results demonstrate that the proposed method achieves an accuracy of 98.42 % for general sow nursing behaviour and 94.37 % for fine-grained nursing behaviour, including nursing with and without the calling-to-nurse sound, and non-nursing behaviours. This fine-grained nursing information can provide a more nuanced understanding of the sow's health and lactation willingness, thereby enhancing management practices in pig farming.
哺乳行为和母猪叫声是评估母猪母性行为和哺乳状况的重要指标。然而,在复杂的室内猪舍中准确识别母猪的这些行为具有挑战性,原因包括光照变化、栏杆障碍物以及其他母猪叫声的干扰。多模态融合(将音频和视觉数据整合在一起)已被证明是在复杂场景中提高准确性和鲁棒性的有效方法。在本研究中,我们设计了一个视听数据采集系统,其中包括一个用于同步采集音频和视频的摄像头,以及一个利用声音传感器追踪声音方向的定制开发的声源定位系统。具体来说,我们提出了一种基于变压器的新型视听多模态融合(TMF)框架,用于识别有或没有母猪叫声的细粒度母猪哺乳行为。最初,我们采用了单模态自我注意增强(USE)模块,利用全局上下文信息增强视频和音频特征。随后,我们开发了视听交互增强(AVIE)模块,利用信息瓶颈原理压缩相关信息并减少噪音。此外,我们还提出了一种自适应动态决策融合策略,通过关注每种模式中最相关的特征来优化模型的性能。最后,我们通过整合音频和融合信息,全面识别了细粒度的哺乳行为,同时结合实时声源定位系统的角度信息,准确判断声音线索是否来自目标母猪。我们的研究结果表明,所提出的方法对一般母猪哺乳行为的准确率达到 98.42%,对细粒度哺乳行为的准确率达到 94.37%,其中包括发出或未发出母猪叫声的哺乳行为以及非哺乳行为。这种精细的哺乳信息可以让人更细致地了解母猪的健康状况和泌乳意愿,从而改进养猪业的管理方法。
{"title":"Transformer-based audio-visual multimodal fusion for fine-grained recognition of individual sow nursing behaviour","authors":"Yuqing Yang ,&nbsp;Chengguo Xu ,&nbsp;Wenhao Hou ,&nbsp;Alan G. McElligott ,&nbsp;Kai Liu ,&nbsp;Yueju Xue","doi":"10.1016/j.aiia.2025.03.006","DOIUrl":"10.1016/j.aiia.2025.03.006","url":null,"abstract":"<div><div>Nursing behaviour and the calling-to-nurse sound are crucial indicators for assessing sow maternal behaviour and nursing status. However, accurately identifying these behaviours for individual sows in complex indoor pig housing is challenging due to factors such as variable lighting, rail obstructions, and interference from other sows' calls. Multimodal fusion, which integrates audio and visual data, has proven to be an effective approach for improving accuracy and robustness in complex scenarios. In this study, we designed an audio-visual data acquisition system that includes a camera for synchronised audio and video capture, along with a custom-developed sound source localisation system that leverages a sound sensor to track sound direction. Specifically, we proposed a novel transformer-based audio-visual multimodal fusion (TMF) framework for recognising fine-grained sow nursing behaviour with or without the calling-to-nurse sound. Initially, a unimodal self-attention enhancement (USE) module was employed to augment video and audio features with global contextual information. Subsequently, we developed an audio-visual interaction enhancement (AVIE) module to compress relevant information and reduce noise using the information bottleneck principle. Moreover, we presented an adaptive dynamic decision fusion strategy to optimise the model's performance by focusing on the most relevant features in each modality. Finally, we comprehensively identified fine-grained nursing behaviours by integrating audio and fused information, while incorporating angle information from the real-time sound source localisation system to accurately determine whether the sound cues originate from the target sow. Our results demonstrate that the proposed method achieves an accuracy of 98.42 % for general sow nursing behaviour and 94.37 % for fine-grained nursing behaviour, including nursing with and without the calling-to-nurse sound, and non-nursing behaviours. This fine-grained nursing information can provide a more nuanced understanding of the sow's health and lactation willingness, thereby enhancing management practices in pig farming.</div></div>","PeriodicalId":52814,"journal":{"name":"Artificial Intelligence in Agriculture","volume":"15 3","pages":"Pages 363-376"},"PeriodicalIF":8.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143835177","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Artificial Intelligence in Agriculture
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1