首页 > 最新文献

Computers and Electronics in Agriculture最新文献

英文 中文
AI-driven analysis of animal cleanliness: A data-fusion model using RGB and thermal imaging 人工智能驱动的动物清洁度分析:使用RGB和热成像的数据融合模型
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2026-01-20 DOI: 10.1016/j.compag.2026.111462
Alberto Carraro , Giulia Bugin , Francesco Marinello, Maddi Aguirrebengoa, Stefano Frattini, Andrea Pezzuolo
Accurate assessment of dairy cow cleanliness is essential for ensuring animal welfare, maintaining udder health, and optimising milk production. Traditional visual inspections are subjective and often fail to distinguish dirt from natural coat patterns, especially in spotted breeds. This research investigates the applicability of a two-stage approach for automated cleanliness evaluation, consisting of (i) semantic segmentation of dirt areas on cow coats and (ii) regression from the resulting masks to numerical cleanliness scores. The first stage was implemented using the U-Net and DeepLabV3 architectures, which were trained on either RGB-only or RGB-Thermal (RGB-T) images. Incorporating thermal information significantly improved segmentation accuracy: U-Net achieved a mean Intersection over Union (mIoU) of 0.5244 on RGB-T images, compared to 0.3537 on RGB images, while DeepLabV3 on RGB-T images reached an mIoU of 0.5049. The second stage compared two regression strategies: multiple linear regression (MLR) on the number of pixels classified as dirt, and convolutional neural networks (CNNs) trained directly on the masks. CNN-based regression consistently outperformed MLR, with the best performance obtained by combining RGB-T segmentation and CNN regression (DeepLabV3 + CNN: MAPE 23.05 %; U-Net + CNN: MAPE 25.24 %). These findings support the feasibility of a two-stage RGB-T-based approach for objective cleanliness evaluation, highlighting the benefits of thermal information for segmentation and CNNs for score prediction.
准确评估奶牛清洁度对确保动物福利、保持乳房健康和优化牛奶产量至关重要。传统的目视检查是主观的,往往不能区分污垢和自然的被毛图案,特别是在斑点品种。本研究探讨了自动清洁度评估的两阶段方法的适用性,包括(i)对奶牛皮毛上污垢区域的语义分割和(ii)从结果掩模到数值清洁度分数的回归。第一阶段使用U-Net和DeepLabV3架构实现,它们在RGB-only或RGB-Thermal (RGB-T)图像上进行训练。结合热信息显著提高了分割精度:U-Net在RGB- t图像上的平均交叉比(Intersection over Union, mIoU)为0.5244,而在RGB图像上为0.3537,而DeepLabV3在RGB- t图像上的mIoU为0.5049。第二阶段比较了两种回归策略:对分类为污垢的像素数量进行多元线性回归(MLR),以及直接在掩模上训练的卷积神经网络(cnn)。基于CNN的回归始终优于MLR,其中RGB-T分割与CNN回归相结合的效果最好(DeepLabV3 + CNN: MAPE 23.05%; U-Net + CNN: MAPE 25.24%)。这些发现支持了基于rgb的两阶段客观清洁度评估方法的可行性,突出了热信息用于分割和cnn用于评分预测的好处。
{"title":"AI-driven analysis of animal cleanliness: A data-fusion model using RGB and thermal imaging","authors":"Alberto Carraro ,&nbsp;Giulia Bugin ,&nbsp;Francesco Marinello,&nbsp;Maddi Aguirrebengoa,&nbsp;Stefano Frattini,&nbsp;Andrea Pezzuolo","doi":"10.1016/j.compag.2026.111462","DOIUrl":"10.1016/j.compag.2026.111462","url":null,"abstract":"<div><div>Accurate assessment of dairy cow cleanliness is essential for ensuring animal welfare, maintaining udder health, and optimising milk production. Traditional visual inspections are subjective and often fail to distinguish dirt from natural coat patterns, especially in spotted breeds. This research investigates the applicability of a two-stage approach for automated cleanliness evaluation, consisting of (i) semantic segmentation of dirt areas on cow coats and (ii) regression from the resulting masks to numerical cleanliness scores. The first stage was implemented using the U-Net and DeepLabV3 architectures, which were trained on either RGB-only or RGB-Thermal (RGB-T) images. Incorporating thermal information significantly improved segmentation accuracy: U-Net achieved a mean Intersection over Union (mIoU) of 0.5244 on RGB-T images, compared to 0.3537 on RGB images, while DeepLabV3 on RGB-T images reached an mIoU of 0.5049. The second stage compared two regression strategies: multiple linear regression (MLR) on the number of pixels classified as dirt, and convolutional neural networks (CNNs) trained directly on the masks. CNN-based regression consistently outperformed MLR, with the best performance obtained by combining RGB-T segmentation and CNN regression (DeepLabV3 + CNN: MAPE 23.05 %; U-Net + CNN: MAPE 25.24 %). These findings support the feasibility of a two-stage RGB-T-based approach for objective cleanliness evaluation, highlighting the benefits of thermal information for segmentation and CNNs for score prediction.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"244 ","pages":"Article 111462"},"PeriodicalIF":8.9,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146025266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient force and stiffness prediction in robotic produce handling with a piezoresistive pressure sensor 用压阻式压力传感器预测机器人产品搬运的有效力和刚度
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2026-01-19 DOI: 10.1016/j.compag.2025.111391
Preston Fairchild, Claudia Chen, Xiaobo Tan
Properly handling delicate produce with robotic manipulators is a major part of the future role of automation in agricultural harvesting and processing. Grasping with the correct amount of force is crucial in not only ensuring proper grip on the object, but also to avoid damaging or bruising the product. In this work, a flexible pressure sensor that is both low cost and easy to fabricate is integrated with robotic grippers for working with produce of varying shapes, sizes, and stiffnesses. The sensor is successfully integrated with both a rigid robotic gripper, as well as a pneumatically actuated soft finger. Furthermore, an algorithm is proposed for accelerated estimation of the steady-state value of the sensor output based on the transient response data, to enable real-time applications. The sensor is shown to be effective in incorporating feedback to correctly grasp objects of unknown sizes and stiffnesses. At the same time, the sensor provides estimates for these values which can be utilized for identification of qualities such as ripeness levels and bruising. It is also shown to be able to provide force feedback for objects of variable stiffnesses. This enables future use not only for produce identification, but also for tasks such as quality control and selective distribution based on ripeness levels.
用机器人正确处理精致的农产品是未来农业收获和加工自动化的重要组成部分。用正确的力量抓握是至关重要的,不仅要确保正确地抓住物体,而且要避免损坏或挫伤产品。在这项工作中,一种低成本且易于制造的柔性压力传感器与机器人抓手集成在一起,用于处理不同形状、尺寸和刚度的产品。该传感器成功地与刚性机器人抓手以及气动驱动的软手指集成在一起。此外,提出了一种基于瞬态响应数据的传感器输出稳态值加速估计算法,以实现实时应用。结果表明,该传感器能够有效地结合反馈信息,正确地抓取未知尺寸和刚度的物体。同时,传感器提供这些值的估计值,这些值可用于识别成熟度和瘀伤等品质。它还显示出能够为变刚度物体提供力反馈。这使得未来不仅可以用于产品识别,还可以用于质量控制和基于成熟度水平的选择性分配等任务。
{"title":"Efficient force and stiffness prediction in robotic produce handling with a piezoresistive pressure sensor","authors":"Preston Fairchild,&nbsp;Claudia Chen,&nbsp;Xiaobo Tan","doi":"10.1016/j.compag.2025.111391","DOIUrl":"10.1016/j.compag.2025.111391","url":null,"abstract":"<div><div>Properly handling delicate produce with robotic manipulators is a major part of the future role of automation in agricultural harvesting and processing. Grasping with the correct amount of force is crucial in not only ensuring proper grip on the object, but also to avoid damaging or bruising the product. In this work, a flexible pressure sensor that is both low cost and easy to fabricate is integrated with robotic grippers for working with produce of varying shapes, sizes, and stiffnesses. The sensor is successfully integrated with both a rigid robotic gripper, as well as a pneumatically actuated soft finger. Furthermore, an algorithm is proposed for accelerated estimation of the steady-state value of the sensor output based on the transient response data, to enable real-time applications. The sensor is shown to be effective in incorporating feedback to correctly grasp objects of unknown sizes and stiffnesses. At the same time, the sensor provides estimates for these values which can be utilized for identification of qualities such as ripeness levels and bruising. It is also shown to be able to provide force feedback for objects of variable stiffnesses. This enables future use not only for produce identification, but also for tasks such as quality control and selective distribution based on ripeness levels.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"244 ","pages":"Article 111391"},"PeriodicalIF":8.9,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146025085","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mechanism-guided deep learning for pest classification in tomato leaves 机制引导的番茄叶片害虫分类深度学习
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2026-01-19 DOI: 10.1016/j.compag.2026.111434
Mingji Wei , Fei Lyu , Shuai Lu , Weijie Liu , Zhaoxuan Fan , Ning Yang , Wenhao Hui
Tomato, a vital global crop, faces severe annual yield losses of 20–40% due to the pest stress, while conventional identification methods often fail to detect early infections before visible symptoms emerge. To overcome the limited sensitivity and generalizability of image-based approaches, this article proposed a tomato pest stress classification method based on a mechanism-guided hybrid network (MGHN) to provide the theoretical foundation for real-time early-warning systems. We first extracted three mechanism-driven features including transient change rates, pulse amplitude/frequency and baseline offset from the dynamics of jasmonic acid (JA) and salicylic acid (SA). These features were then fused with a 1D-CNN to classify the three pest types including piercing-sucking, chewing and crawling-feeding pests. Results show that our proposed MGHN method achieves 94.8% average accuracy on 135 independent test samples, with significantly optimized recognition for piercing-sucking pests (95.56% F1), chewing pests (94.50% F1), and crawling-feeding pests (94.37% F1). Comparative analysis against traditional 1D-CNN and a mechanism-only model demonstrates that MGHN outperformed 1D-CNN (91.1%) and the mechanism-only model (89.6%) by 3.7% and 5.2%, respectively. This research can establish a theoretical foundation for real-time early pest stress warning systems of crops using wearable sensors.
西红柿是一种重要的全球作物,由于虫害胁迫,每年面临20-40%的严重产量损失,而传统的鉴定方法往往无法在出现明显症状之前发现早期感染。为了克服基于图像的方法灵敏度和泛化性的局限性,本文提出了一种基于机制导向混合网络(MGHN)的番茄有害生物胁迫分类方法,为实时预警系统提供理论基础。我们首先从茉莉酸(JA)和水杨酸(SA)的动力学中提取了三个机制驱动的特征,包括瞬态变化率、脉冲幅度/频率和基线偏移。然后将这些特征与1D-CNN融合,以分类三种害虫类型,包括刺吸,咀嚼和爬食害虫。结果表明,本文提出的MGHN方法在135个独立测试样本上的平均准确率达到94.8%,其中对刺吸型害虫(95.56% F1)、咀嚼型害虫(94.50% F1)和爬食型害虫(94.37% F1)的识别效果显著。与传统1D-CNN和纯机制模型的对比分析表明,MGHN分别比1D-CNN(91.1%)和纯机制模型(89.6%)高出3.7%和5.2%。本研究可为基于可穿戴传感器的作物病虫害实时预警系统奠定理论基础。
{"title":"Mechanism-guided deep learning for pest classification in tomato leaves","authors":"Mingji Wei ,&nbsp;Fei Lyu ,&nbsp;Shuai Lu ,&nbsp;Weijie Liu ,&nbsp;Zhaoxuan Fan ,&nbsp;Ning Yang ,&nbsp;Wenhao Hui","doi":"10.1016/j.compag.2026.111434","DOIUrl":"10.1016/j.compag.2026.111434","url":null,"abstract":"<div><div>Tomato, a vital global crop, faces severe annual yield losses of 20–40% due to the pest stress, while conventional identification methods often fail to detect early infections before visible symptoms emerge. To overcome the limited sensitivity and generalizability of image-based approaches, this article proposed a tomato pest stress classification method based on a mechanism-guided hybrid network (MGHN) to provide the theoretical foundation for real-time early-warning systems. We first extracted three mechanism-driven features including transient change rates, pulse amplitude/frequency and baseline offset from the dynamics of jasmonic acid (JA) and salicylic acid (SA). These features were then fused with a 1D-CNN to classify the three pest types including piercing-sucking, chewing and crawling-feeding pests. Results show that our proposed MGHN method achieves 94.8% average accuracy on 135 independent test samples, with significantly optimized recognition for piercing-sucking pests (95.56% <em>F</em>1), chewing pests (94.50% <em>F</em>1), and crawling-feeding pests (94.37% <em>F</em>1). Comparative analysis against traditional 1D-CNN and a mechanism-only model demonstrates that MGHN outperformed 1D-CNN (91.1%) and the mechanism-only model (89.6%) by 3.7% and 5.2%, respectively. This research can establish a theoretical foundation for real-time early pest stress warning systems of crops using wearable sensors.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"244 ","pages":"Article 111434"},"PeriodicalIF":8.9,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146025088","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Joint cascaded 3DCNN and SDTA encoding for tree species classification using UAV-based hyperspectral image in mining areas 3DCNN与SDTA联合级联编码在矿区无人机高光谱影像树种分类中的应用
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2026-01-19 DOI: 10.1016/j.compag.2026.111438
Yuncheng Deng , Suling He , Jinliang Wang , Jianpeng Zhang , Bangjin Yi , Congtao Hu , Yanling Jiang , An Chen
Reforestation is crucial for ecological restoration in mining areas, and precise classification of tree species is an important prerequisite for evaluating the effectiveness of ecological restoration. Unmanned aerial vehicles (UAVs) hyperspectral imagery enables highly accurate classification of tree species, leveraging its superior spectral and spatial resolutions. However, the challenge of dimensionality escalation, caused by hundreds of spectral features across numerous bands, introduces new hurdles for conventional classification methods. Deep learning provides a new solution for automatic feature extraction and tree classification and mapping. However, there are problems such as single scale of extracted features, static model inference, and weak model generalization ability. Especially in the ecological restoration area of the mining area, where the terrain environment is complex and tree species of multiple forest age levels coexist, and the classification accuracy needs to be improved. Therefore, this study uses UAV-based hyperspectral imagery acquired from the Jianshan mining area in Kunming, Yunnan Province, as the primary data source, and proposes a novel three-dimensional convolutional neural network (SDTA-3DCNN) incorporating a separable depth transposed attention mechanism. The model enhances the extraction of high-dimensional sparse spectral features through a cascaded 3DCNN architecture, while effectively resolving complex spectral patterns by leveraging an attention mechanism to fuse local features with high-level global representations. The final tree classification accuracy reaches F1-score = 0.9814, OA = 0.9903 and Kappa = 0.9868. The proposed method is also well applied in other tree classification and crop classification scenarios, with OA and Kappa of 0.9871 and 0.9688 in other tree classification scenarios, respectively, and the highest classification accuracy in crop classification scenarios reaches OA = 0.9712, Kappa = 0.9635. The method can provide scientific basis and technical support for the fine classification of tree species in mining areas, the monitoring of ecological restoration status in mining areas, the evaluation of effects, and the decision-making.
重新造林是矿区生态恢复的关键,而精确的树种分类是评价矿区生态恢复效果的重要前提。无人机(uav)的高光谱图像能够利用其优越的光谱和空间分辨率对树种进行高度精确的分类。然而,由于多个波段的数百个光谱特征导致了维度升级的挑战,这给传统的分类方法带来了新的障碍。深度学习为自动特征提取和树分类映射提供了新的解决方案。但存在特征提取尺度单一、模型推理静态、模型泛化能力弱等问题。特别是矿区生态恢复区地形环境复杂,多种林龄水平树种并存,分类精度有待提高。因此,本研究以云南昆明尖山矿区的无人机高光谱影像为主要数据源,提出了一种包含可分离深度转置注意机制的三维卷积神经网络(SDTA-3DCNN)。该模型通过级联3DCNN架构增强了高维稀疏光谱特征的提取,同时利用注意力机制将局部特征与高级全局表征融合,有效地解决了复杂的光谱模式。最终的树分类准确率达到F1-score = 0.9814, OA = 0.9903, Kappa = 0.9868。该方法在其他树木分类和作物分类场景中也有很好的应用,其他树木分类场景的OA和Kappa分别为0.9871和0.9688,作物分类场景的分类准确率最高,OA = 0.9712, Kappa = 0.9635。该方法可为矿区树种的精细分类、矿区生态修复状况监测、效果评价和决策提供科学依据和技术支持。
{"title":"Joint cascaded 3DCNN and SDTA encoding for tree species classification using UAV-based hyperspectral image in mining areas","authors":"Yuncheng Deng ,&nbsp;Suling He ,&nbsp;Jinliang Wang ,&nbsp;Jianpeng Zhang ,&nbsp;Bangjin Yi ,&nbsp;Congtao Hu ,&nbsp;Yanling Jiang ,&nbsp;An Chen","doi":"10.1016/j.compag.2026.111438","DOIUrl":"10.1016/j.compag.2026.111438","url":null,"abstract":"<div><div>Reforestation is crucial for ecological restoration in mining areas, and precise classification of tree species is an important prerequisite for evaluating the effectiveness of ecological restoration. Unmanned aerial vehicles (UAVs) hyperspectral imagery enables highly accurate classification of tree species, leveraging its superior spectral and spatial resolutions. However, the challenge of dimensionality escalation, caused by hundreds of spectral features across numerous bands, introduces new hurdles for conventional classification methods. Deep learning provides a new solution for automatic feature extraction and tree classification and mapping. However, there are problems such as single scale of extracted features, static model inference, and weak model generalization ability. Especially in the ecological restoration area of the mining area, where the terrain environment is complex and tree species of multiple forest age levels coexist, and the classification accuracy needs to be improved. Therefore, this study uses UAV-based hyperspectral imagery acquired from the Jianshan mining area in Kunming, Yunnan Province, as the primary data source, and proposes a novel three-dimensional convolutional neural network (SDTA-3DCNN) incorporating a separable depth transposed attention mechanism. The model enhances the extraction of high-dimensional sparse spectral features through a cascaded 3DCNN architecture, while effectively resolving complex spectral patterns by leveraging an attention mechanism to fuse local features with high-level global representations. The final tree classification accuracy reaches F1-score = 0.9814, OA = 0.9903 and Kappa = 0.9868. The proposed method is also well applied in other tree classification and crop classification scenarios, with OA and Kappa of 0.9871 and 0.9688 in other tree classification scenarios, respectively, and the highest classification accuracy in crop classification scenarios reaches OA = 0.9712, Kappa = 0.9635. The method can provide scientific basis and technical support for the fine classification of tree species in mining areas, the monitoring of ecological restoration status in mining areas, the evaluation of effects, and the decision-making.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"244 ","pages":"Article 111438"},"PeriodicalIF":8.9,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146025089","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development and performance evaluation of an automatic control system for sugarcane harvester extractor 甘蔗收获机抽提机自动控制系统的研制与性能评价
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2026-01-19 DOI: 10.1016/j.compag.2026.111425
Baocheng Zhou , Shaochun Ma , Wenzhi Li , Jinzhi Ma , Yansu Xie , Sha Yang
Real-time adjustment of extractor speed according to feed rate is essential to reduce impurity content and cane loss in mechanized sugarcane harvesting. An automatic control system for sugarcane harvester extractor was developed in this study aiming to achieve dynamic matching between speed and feed rate, thereby reducing impurity content and cane loss during harvesting. An optimal control strategy between feed rate and rotational speed was established using impurity content and cane loss as indicators. A variable universe fuzzy multi-parameter adaptive PID (VUFMA-PID) control method was proposed and modeled in Simulink. Compared with conventional PID and fuzzy PID, the VUFMA-PID achieved the shortest steady-state response time, 0.32 s and 0.26 s faster than PID and fuzzy PID, with both steady-state error and maximum overshoot reduced to zero. Field experiments were conducted under different feed rate fluctuation orders, with fixed extractor speed and manual adjustment speed based on operator experience used as control groups. The results indicated that, compared to manual and constant mode, the average power consumption of the automatic control mode was reduced by 17.44 % and 30.40 % respectively. The average impurity content was 4.00 %, which decreased by 23.58 % and 10.71 %. The average cane loss was 1.89 %, which decreased by 25.01 % and 28.52 %. The developed automatic control system effectively adapts to varying feed rates and significantly improves harvesting quality. It provides a feasible solution and theoretical support for intelligent control in mechanized sugarcane harvesting.
在甘蔗机械化采收过程中,根据进料速率实时调节抽提器转速是降低杂质含量和减少甘蔗损失的关键。本研究开发了一种甘蔗收获机抽采机自动控制系统,实现速度与进料速度的动态匹配,从而减少收获过程中的杂质含量和甘蔗损失。以杂质含量和甘蔗损失为指标,建立了进料速率与转速的最优控制策略。提出了一种变域模糊多参数自适应PID (VUFMA-PID)控制方法,并在Simulink中建模。与传统PID和模糊PID相比,VUFMA-PID的稳态响应时间最短,分别比PID和模糊PID快0.32 s和0.26 s,且稳态误差和最大超调量均降至零。在不同进料速率波动顺序下进行现场实验,以固定的提取速度和根据操作人员经验手动调节速度作为对照组。结果表明,与手动和恒定模式相比,自动控制模式的平均功耗分别降低了17.44%和30.40%。平均杂质含量为4.00%,分别下降23.58%和10.71%。甘蔗平均损失率为1.89%,比上年分别下降25.01%和28.52%。开发的自动控制系统能有效适应不同进料速率,显著提高收获质量。为甘蔗机械化采收智能化控制提供了可行方案和理论支持。
{"title":"Development and performance evaluation of an automatic control system for sugarcane harvester extractor","authors":"Baocheng Zhou ,&nbsp;Shaochun Ma ,&nbsp;Wenzhi Li ,&nbsp;Jinzhi Ma ,&nbsp;Yansu Xie ,&nbsp;Sha Yang","doi":"10.1016/j.compag.2026.111425","DOIUrl":"10.1016/j.compag.2026.111425","url":null,"abstract":"<div><div>Real-time adjustment of extractor speed according to feed rate is essential to reduce impurity content and cane loss in mechanized sugarcane harvesting. An automatic control system for sugarcane harvester extractor was developed in this study aiming to achieve dynamic matching between speed and feed rate, thereby reducing impurity content and cane loss during harvesting. An optimal control strategy between feed rate and rotational speed was established using impurity content and cane loss as indicators. A variable universe fuzzy multi-parameter adaptive PID (VUFMA-PID) control method was proposed and modeled in Simulink. Compared with conventional PID and fuzzy PID, the VUFMA-PID achieved the shortest steady-state response time, 0.32 s and 0.26 s faster than PID and fuzzy PID, with both steady-state error and maximum overshoot reduced to zero. Field experiments were conducted under different feed rate fluctuation orders, with fixed extractor speed and manual adjustment speed based on operator experience used as control groups. The results indicated that, compared to manual and constant mode, the average power consumption of the automatic control mode was reduced by 17.44 % and 30.40 % respectively. The average impurity content was 4.00 %, which decreased by 23.58 % and 10.71 %. The average cane loss was 1.89 %, which decreased by 25.01 % and 28.52 %. The developed automatic control system effectively adapts to varying feed rates and significantly improves harvesting quality. It provides a feasible solution and theoretical support for intelligent control in mechanized sugarcane harvesting.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"244 ","pages":"Article 111425"},"PeriodicalIF":8.9,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146025087","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Yield forecasting in maize: Performance and limits of unmanned aerial vehicle and PlanetScope remote sensing across multiple growth cycles 玉米产量预测:无人机和PlanetScope遥感跨多个生长周期的性能和局限性
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2026-01-19 DOI: 10.1016/j.compag.2026.111445
Philippe Vigneault , Samuel de la Sablonnière , Arianne Deshaies , Kosal Khun , Joel Lafond-Lapalme , Louis Longchamps , Étienne Lord
This study addresses the challenge of forecasting maize yield in southeastern Quebec by comparing weekly Unmanned Aerial Vehicle (UAV) and PlanetScope satellite imagery throughout the cropping season and across diverse growing conditions. Using five nitrogen treatments over three years with two sowing windows each year to generate variability within the dataset, eleven vegetation indices were evaluated to identify the best-performing indices and the optimal forecasting window. Indices were interpolated using curve fitting to enable evaluation at any stage of the growing season. Cross-validation simulated real-world application by excluding entire sowing events during model testing. Using a linear regression approach, results demonstrate that indices combining green and near-infrared bands (Green Normalized Difference Vegetation Index [GNDVI] and Chlorophyll Index Green [CIG]) exhibit superior forecasting potential compared to red-near-infrared (like NDVI) and RGB-based indices (like NGRDI). The optimal forecast window occurs during early grain filling (R2-R3 stages, around 2300 Crop Heat Units [CHU]), achieving Root Mean Square Coefficient of Variation (RMSCV) values of 12.51 % for UAVs and 15.28 % for PlanetScope. While PlanetScope maximum performance approached UAV capabilities, results showed a CHU range between 200 and 400 in the effective forecasting period (RMSCV < 20 %) compared to 850 to 1450 for UAV. For PlanetScope, adding multiple indices marginally improved precision, slightly reduced forecasting window and reduced model transferability. The analysis revealed weak correlations between indices and yield during early vegetative and senescence phases, indicating limited potential for enabling timely in-season management interventions. This study established UAV-based models as a reference point for assessing the limitations of satellite-derived forecasts.
本研究通过比较整个种植季节和不同生长条件下的每周无人机(UAV)和PlanetScope卫星图像,解决了预测魁北克东南部玉米产量的挑战。采用3年5个氮肥处理,每年2个播种窗,在数据集中产生变异,对11个植被指数进行评估,以确定表现最佳的指数和最佳预测窗口。利用曲线拟合对指标进行插值,以便在生长季节的任何阶段进行评价。交叉验证通过在模型测试期间排除整个播种事件来模拟实际应用。利用线性回归方法,研究结果表明,绿色和近红外波段组合的指数(绿色归一化植被指数[GNDVI]和叶绿素指数green [CIG])比基于红-近红外(如NDVI)和基于rgb的指数(如NGRDI)具有更强的预测潜力。最佳预测窗口出现在灌浆早期(R2-R3阶段,约2300作物热量单位[CHU]),实现无人机的均方根变异系数(RMSCV)值为12.51%,PlanetScope为15.28%。虽然PlanetScope的最大性能接近无人机的能力,但结果显示,在有效预测期内,CHU的范围在200到400之间(RMSCV < 20%),而无人机的范围为850到1450。对于PlanetScope,增加多个指标略微提高了精度,略微减少了预测窗口,降低了模型的可转移性。分析显示,在营养早期和衰老阶段,指数与产量之间的相关性较弱,表明及时实施季节性管理干预的潜力有限。本研究建立了基于无人机的模型,作为评估卫星衍生预报局限性的参考点。
{"title":"Yield forecasting in maize: Performance and limits of unmanned aerial vehicle and PlanetScope remote sensing across multiple growth cycles","authors":"Philippe Vigneault ,&nbsp;Samuel de la Sablonnière ,&nbsp;Arianne Deshaies ,&nbsp;Kosal Khun ,&nbsp;Joel Lafond-Lapalme ,&nbsp;Louis Longchamps ,&nbsp;Étienne Lord","doi":"10.1016/j.compag.2026.111445","DOIUrl":"10.1016/j.compag.2026.111445","url":null,"abstract":"<div><div>This study addresses the challenge of forecasting maize yield in southeastern Quebec by comparing weekly Unmanned Aerial Vehicle (UAV) and PlanetScope satellite imagery throughout the cropping season and across diverse growing conditions. Using five nitrogen treatments over three years with two sowing windows each year to generate variability within the dataset, eleven vegetation indices were evaluated to identify the best-performing indices and the optimal forecasting window. Indices were interpolated using curve fitting to enable evaluation at any stage of the growing season. Cross-validation simulated real-world application by excluding entire sowing events during model testing. Using a linear regression approach, results demonstrate that indices combining green and near-infrared bands (Green Normalized Difference Vegetation Index [GNDVI] and Chlorophyll Index Green [CIG]) exhibit superior forecasting potential compared to red-near-infrared (like NDVI) and RGB-based indices (like NGRDI). The optimal forecast window occurs during early grain filling (R2-R3 stages, around 2300 Crop Heat Units [CHU]), achieving Root Mean Square Coefficient of Variation (RMSCV) values of 12.51 % for UAVs and 15.28 % for PlanetScope. While PlanetScope maximum performance approached UAV capabilities, results showed a CHU range between 200 and 400 in the effective forecasting period (RMSCV &lt; 20 %) compared to 850 to 1450 for UAV. For PlanetScope, adding multiple indices marginally improved precision, slightly reduced forecasting window and reduced model transferability. The analysis revealed weak correlations between indices and yield during early vegetative and senescence phases, indicating limited potential for enabling timely in-season management interventions. This study established UAV-based models as a reference point for assessing the limitations of satellite-derived forecasts.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"244 ","pages":"Article 111445"},"PeriodicalIF":8.9,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146025265","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Beyond Color: Advanced RGB-D data augmentation for robust semantic segmentation in crop farming scenes 超越颜色:先进的RGB-D数据增强,用于农作物种植场景的鲁棒语义分割
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2026-01-18 DOI: 10.1016/j.compag.2026.111432
Florian Kitzler , Alexander Bauer , Viktoria Kruder-Motsch
The emergence of smart farming in recent years has substantially increased the importance of artificial vision systems in crop production. Data augmentation is essential for developing robust semantic segmentation models when dealing with small datasets, such as in selective weed control. Due to advances in multi-modal data fusion, RGB-D image datasets contribute substantially to improve model performance. However, most data augmentation techniques primarily modify the color channels, often neglecting the depth channel. Addressing this gap, we introduce three methods for augmenting RGB-D images: RGB-D-Aug, Recompose3D, and Compose3D. We conducted experiments utilizing a multi-modal fusion network tailored for semantic segmentation of different plant species, namely ESANet. RGB-D-Aug introduces artificial depth sensor noise in addition to commonly used geometric transformations and color variations. Recompose3D and Compose3D generate augmented RGB-D images and corresponding ground-truth labels by composing background images and a set of foreground plant snippets. Recompose3D rearranges plants from a given training image, while Compose3D employs all plant snippets available in the training dataset. In our experiments designed to evaluate generalization performance, we tested our three methods and compared them not only to the augmentation technique used in ESANet, which consists of geometric transformations and color channel variations, but also to an extended version of the Copy-Paste method, an image composition technique originally introduced for RGB images. All three of our proposed methods outperformed the ESANet augmentation. The image composition methods, Copy-Paste, Recompose3D, and Compose3D, performed significantly better, with Compose3D achieving the highest generalization performance of all methods tested. In addition to improving model robustness, Compose3D allows the creation of realistic agronomic image scenes. Our research is an important step towards developing robust and generalizable models for different applications in arable farming.
近年来智能农业的出现大大增加了人工视觉系统在作物生产中的重要性。当处理小数据集(如选择性杂草控制)时,数据增强对于开发健壮的语义分割模型至关重要。由于多模态数据融合的进步,RGB-D图像数据集对提高模型性能有很大贡献。然而,大多数数据增强技术主要是修改颜色通道,往往忽略了深度通道。为了解决这一问题,我们介绍了三种增强RGB-D图像的方法:RGB-D- aug、Recompose3D和Compose3D。我们利用一个多模态融合网络进行了实验,该网络是为不同植物物种的语义分割量身定制的,即ESANet。RGB-D-Aug除了常用的几何变换和颜色变化外,还引入了人工深度传感器噪声。Recompose3D和Compose3D通过合成背景图像和一组前景植物片段来生成增强的RGB-D图像和相应的真地标签。Recompose3D从给定的训练图像中重新排列植物,而Compose3D使用训练数据集中所有可用的植物片段。在我们旨在评估泛化性能的实验中,我们测试了我们的三种方法,并将它们不仅与ESANet中使用的增强技术(由几何变换和颜色通道变化组成)进行了比较,还与复制-粘贴方法的扩展版本进行了比较,复制-粘贴方法是一种最初为RGB图像引入的图像合成技术。我们提出的所有三种方法都优于ESANet增强。图像合成方法Copy-Paste、Recompose3D和Compose3D表现明显更好,其中Compose3D在所有测试方法中实现了最高的泛化性能。除了提高模型鲁棒性,Compose3D允许创建逼真的农艺学图像场景。我们的研究是朝着开发健壮的和可推广的模型在耕地农业中的不同应用迈出的重要一步。
{"title":"Beyond Color: Advanced RGB-D data augmentation for robust semantic segmentation in crop farming scenes","authors":"Florian Kitzler ,&nbsp;Alexander Bauer ,&nbsp;Viktoria Kruder-Motsch","doi":"10.1016/j.compag.2026.111432","DOIUrl":"10.1016/j.compag.2026.111432","url":null,"abstract":"<div><div>The emergence of smart farming in recent years has substantially increased the importance of artificial vision systems in crop production. Data augmentation is essential for developing robust semantic segmentation models when dealing with small datasets, such as in selective weed control. Due to advances in multi-modal data fusion, RGB-D image datasets contribute substantially to improve model performance. However, most data augmentation techniques primarily modify the color channels, often neglecting the depth channel. Addressing this gap, we introduce three methods for augmenting RGB-D images: RGB-D-Aug, Recompose3D, and Compose3D. We conducted experiments utilizing a multi-modal fusion network tailored for semantic segmentation of different plant species, namely ESANet. RGB-D-Aug introduces artificial depth sensor noise in addition to commonly used geometric transformations and color variations. Recompose3D and Compose3D generate augmented RGB-D images and corresponding ground-truth labels by composing background images and a set of foreground plant snippets. Recompose3D rearranges plants from a given training image, while Compose3D employs all plant snippets available in the training dataset. In our experiments designed to evaluate generalization performance, we tested our three methods and compared them not only to the augmentation technique used in ESANet, which consists of geometric transformations and color channel variations, but also to an extended version of the Copy-Paste method, an image composition technique originally introduced for RGB images. All three of our proposed methods outperformed the ESANet augmentation. The image composition methods, Copy-Paste, Recompose3D, and Compose3D, performed significantly better, with Compose3D achieving the highest generalization performance of all methods tested. In addition to improving model robustness, Compose3D allows the creation of realistic agronomic image scenes. Our research is an important step towards developing robust and generalizable models for different applications in arable farming.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"244 ","pages":"Article 111432"},"PeriodicalIF":8.9,"publicationDate":"2026-01-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146025086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Magnetoelastomer-based grain flow sensor for combine harvesters 用于联合收割机的磁弹性体谷物流量传感器
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2026-01-18 DOI: 10.1016/j.compag.2026.111447
Xin-Dong Ni , Mao-Lin Li , Fang-Lei Li , Zi-Xuan Dai , Chun-Xiao Xing , Feng Wang , Yan-Xin Yin , Du Chen , Zhi-Zhu He
Accurate and stable grain flow monitoring plays a critical role in yield estimation and closed-loop control of combine harvesters. To address the limitations of existing flow sensors under dynamic and noisy field conditions, this study proposes a novel grain flow sensing method based on a magnetoelastomer–Hall array structure. A multilayer flexible composite structure, comprising a NdFeB magnetic film and a PDMS elastomeric substrate, was developed to enable direct, in-situ transduction of grain flow impact into magnetic field variations. The sensor was structurally designed using discrete element method (DEM) simulations to align with the spatial impact distribution patterns of free-falling grains. A 3 × 3 magnetoelastomer array was arranged to capture the spatial characteristics of grain impacts. And a maximum-response extraction strategy was adopted to enhance signal robustness under heterogeneous grain flow. Laboratory tests confirmed that 3 × 3 array architecture preserves the mechanical compliance and sensitivity of the magnetoelastomer units without performance trade-offs. Grain flow detection performance was evaluated on a grain flow test bench and experimental results revealed a strong linear relationship between sensor output and actual grain flow (R2=0.98), with a mean yield estimation error of 6.07%. Therefore, the proposed sensing system provides a flexible, sensitive, and reliable solution for real-time grain flow monitoring, and establishes a foundation for introducing flexible sensing into next-generation intelligent agricultural machinery.
准确稳定的粮流监测对联合收割机产量估算和闭环控制具有重要意义。为了解决现有流量传感器在动态和噪声场条件下的局限性,本研究提出了一种基于磁弹性体-霍尔阵列结构的新型颗粒流量传感方法。研究人员开发了一种多层柔性复合结构,包括钕铁硼磁性薄膜和PDMS弹性衬底,可以将晶粒流动影响直接转化为磁场变化。为了适应自由落体颗粒的空间冲击分布规律,采用离散元法(DEM)对传感器进行了结构设计。采用3 × 3磁弹性体阵列捕捉颗粒碰撞的空间特征。采用最大响应提取策略,增强了非均匀颗粒流条件下信号的鲁棒性。实验室测试证实,3 × 3阵列结构保留了磁弹性体单元的机械顺应性和灵敏度,而没有性能折衷。在粮流试验台上对粮流检测性能进行了评价,实验结果表明,传感器输出与实际粮流之间存在较强的线性关系(R2=0.98),平均产量估计误差为6.07%。因此,该传感系统为粮食流量实时监测提供了灵活、灵敏、可靠的解决方案,为将柔性传感引入下一代智能农业机械奠定了基础。
{"title":"Magnetoelastomer-based grain flow sensor for combine harvesters","authors":"Xin-Dong Ni ,&nbsp;Mao-Lin Li ,&nbsp;Fang-Lei Li ,&nbsp;Zi-Xuan Dai ,&nbsp;Chun-Xiao Xing ,&nbsp;Feng Wang ,&nbsp;Yan-Xin Yin ,&nbsp;Du Chen ,&nbsp;Zhi-Zhu He","doi":"10.1016/j.compag.2026.111447","DOIUrl":"10.1016/j.compag.2026.111447","url":null,"abstract":"<div><div>Accurate and stable grain flow monitoring plays a critical role in yield estimation and closed-loop control of combine harvesters. To address the limitations of existing flow sensors under dynamic and noisy field conditions, this study proposes a novel grain flow sensing method based on a magnetoelastomer–Hall array structure. A multilayer flexible composite structure, comprising a NdFeB magnetic film and a PDMS elastomeric substrate, was developed to enable direct, in-situ transduction of grain flow impact into magnetic field variations. The sensor was structurally designed using discrete element method (DEM) simulations to align with the spatial impact distribution patterns of free-falling grains. A 3 × 3 magnetoelastomer array was arranged to capture the spatial characteristics of grain impacts. And a maximum-response extraction strategy was adopted to enhance signal robustness under heterogeneous grain flow. Laboratory tests confirmed that 3 × 3 array architecture preserves the mechanical compliance and sensitivity of the magnetoelastomer units without performance trade-offs. Grain flow detection performance was evaluated on a grain flow test bench and experimental results revealed a strong linear relationship between sensor output and actual grain flow (<span><math><mrow><msup><mrow><mi>R</mi></mrow><mn>2</mn></msup><mo>=</mo><mtext>0.98</mtext></mrow></math></span>), with a mean yield estimation error of 6.07%. Therefore, the proposed sensing system provides a flexible, sensitive, and reliable solution for real-time grain flow monitoring, and establishes a foundation for introducing flexible sensing into next-generation intelligent agricultural machinery.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"244 ","pages":"Article 111447"},"PeriodicalIF":8.9,"publicationDate":"2026-01-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146025152","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Chrysanthemum classification method via multi-stream deep color space feature fusion 基于多流深颜色空间特征融合的菊花分类方法
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2026-01-18 DOI: 10.1016/j.compag.2026.111455
Jian Jiang , Xichen Yang , Hui Yan , Jia Liu , Yifan Chen , Zhongyuan Mao , Tianshu Wang
Chrysanthemum is a traditional Chinese medicinal herb that is widely employed in various medical applications. The medicinal value of chrysanthemums varies among different types, and this value is highly correlated with their specific classification. Therefore, the classification of chrysanthemums is essential for ensuring their proper medicinal use. However, existing traditional classification methods are often time-consuming and costly, making them less suitable for practical applications. To overcome these limitations, a novel chrysanthemum classification method based on multi-stream deep color space feature fusion is proposed. Firstly, chrysanthemum images are transformed from the RGB color space to the HSV and LAB color spaces. And the classification-aware features are extracted from the H, S, and L channels, respectively. Secondly, a multi-stream deep network is designed, which employing both 1D and 2D networks. The 1D network focus on further analyzing the features from the H, S, and L channels. And, the 2D network extracts deep classification-aware features from the original image. Thirdly, an effective feature fusion method is designed to descript the characteristics of different chrysanthemums more efficiently, which take both inter-layer interaction and inter-path interaction into consideration. Finally, all the features extracted from different streams of the network are fused to gain the deep color space feature. The performance comparisons are conducted on the dataset which contain 4276 real chrysanthemum images of 18 categories. Experimental results demonstrate that the proposed method is more accurate and stable than tested methods, achieving an accuracy of 95.45%, which is approximately 2.07% higher than the best tested method.
菊花是一种传统的中草药,广泛应用于各种医疗用途。菊花的药用价值因品种而异,其药用价值与其具体分类高度相关。因此,菊花的分类对于确保其适当的药用是必不可少的。然而,现有的传统分类方法往往耗时且成本高,不太适合实际应用。为了克服这些局限性,提出了一种基于多流深颜色空间特征融合的菊花分类方法。首先,将菊花图像从RGB色彩空间转换为HSV和LAB色彩空间。分别从H、S、L通道提取分类感知特征。其次,设计了一种采用一维和二维网络的多流深度网络。一维网络侧重于进一步分析H、S和L通道的特征。2D网络从原始图像中提取深度分类感知特征。第三,设计了一种有效的特征融合方法,同时考虑了层间和路径间的相互作用,能够更有效地描述不同菊花的特征。最后,对网络中不同流提取的所有特征进行融合,得到深颜色空间特征。在包含18个类别4276张真实菊花图像的数据集上进行性能比较。实验结果表明,该方法的准确度和稳定性优于现有测试方法,准确率为95.45%,比最佳测试方法提高了约2.07%。
{"title":"Chrysanthemum classification method via multi-stream deep color space feature fusion","authors":"Jian Jiang ,&nbsp;Xichen Yang ,&nbsp;Hui Yan ,&nbsp;Jia Liu ,&nbsp;Yifan Chen ,&nbsp;Zhongyuan Mao ,&nbsp;Tianshu Wang","doi":"10.1016/j.compag.2026.111455","DOIUrl":"10.1016/j.compag.2026.111455","url":null,"abstract":"<div><div>Chrysanthemum is a traditional Chinese medicinal herb that is widely employed in various medical applications. The medicinal value of chrysanthemums varies among different types, and this value is highly correlated with their specific classification. Therefore, the classification of chrysanthemums is essential for ensuring their proper medicinal use. However, existing traditional classification methods are often time-consuming and costly, making them less suitable for practical applications. To overcome these limitations, a novel chrysanthemum classification method based on multi-stream deep color space feature fusion is proposed. Firstly, chrysanthemum images are transformed from the RGB color space to the HSV and LAB color spaces. And the classification-aware features are extracted from the H, S, and L channels, respectively. Secondly, a multi-stream deep network is designed, which employing both 1D and 2D networks. The 1D network focus on further analyzing the features from the H, S, and L channels. And, the 2D network extracts deep classification-aware features from the original image. Thirdly, an effective feature fusion method is designed to descript the characteristics of different chrysanthemums more efficiently, which take both inter-layer interaction and inter-path interaction into consideration. Finally, all the features extracted from different streams of the network are fused to gain the deep color space feature. The performance comparisons are conducted on the dataset which contain 4276 real chrysanthemum images of 18 categories. Experimental results demonstrate that the proposed method is more accurate and stable than tested methods, achieving an accuracy of 95.45%, which is approximately 2.07% higher than the best tested method.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"244 ","pages":"Article 111455"},"PeriodicalIF":8.9,"publicationDate":"2026-01-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146025084","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Discriminative feature representations and heterogeneous fusion for plant leaf recognition 植物叶片识别的判别特征表示与异质融合
IF 8.9 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2026-01-17 DOI: 10.1016/j.compag.2026.111431
Mengjie Ye , Yong Cheng , De Yu , Yongqi Yuan , Ge Jin , Huan Wang
Effective feature representation and heterogeneous fusion are essential for plant leaf recognition. However, existing methods have several limitations, such as insufficient comprehensiveness and distinctiveness in feature representation, as well as a lack of full consideration for the compatibility and complementarity in heterogeneous fusion. In the end, we propose a discriminative shape representation named the bag of multiscale curvature angle cuts (BMCAC) to capture fine curvature and spatial distribution characteristics, an advanced deep representation called the progressive salient deep representation (PSDR) to fully exploit deep convolutional features, and an effective fusion framework termed the K-weighted shape and deep feature fusion (KWFF) to aggregate the local context and global importance of heterogeneous features. Specifically, BMCAC is derived from the curvature angle cuts (CAC), multiscale analysis, and the bag of visual words (BoVW) model; PSDR is constructed by applying progressive downsampling and hierarchical pooling operations to deep convolutional features; and KWFF is developed by encoding neighboring information using homogeneous distance measures while incorporating globally weighted contributions from heterogeneous distance measures. Extensive experiments on four well-known benchmark leaf datasets demonstrate that the proposed shape and deep representations can efficiently extract leaf image features, and the fusion framework can effectively integrate heterogeneous features, outperforming state-of-the-art methods. The source code is available at https://github.com/Mumuxi1123/BMCAC_PSDR_KWFF.
有效的特征表示和异构融合是植物叶片识别的关键。然而,现有的方法存在一些局限性,如特征表示的全面性和独特性不足,以及对异构融合的兼容性和互补性考虑不足。最后,我们提出了一种判别形状表示(BMCAC)来捕捉精细的曲率和空间分布特征,一种先进的深度表示(PSDR)来充分利用深度卷积特征,一种称为k加权形状和深度特征融合(KWFF)的有效融合框架来聚合异构特征的局部背景和全局重要性。具体来说,BMCAC是由曲率角切割(CAC)、多尺度分析和视觉词包(BoVW)模型推导而来;PSDR是通过对深度卷积特征应用渐进降采样和分层池化操作来构建的;KWFF是通过使用同质距离度量编码相邻信息,同时结合异构距离度量的全局加权贡献来开发的。在4个知名基准叶片数据集上的大量实验表明,所提出的形状表征和深度表征可以有效地提取叶片图像特征,融合框架可以有效地整合异构特征,优于现有方法。源代码可从https://github.com/Mumuxi1123/BMCAC_PSDR_KWFF获得。
{"title":"Discriminative feature representations and heterogeneous fusion for plant leaf recognition","authors":"Mengjie Ye ,&nbsp;Yong Cheng ,&nbsp;De Yu ,&nbsp;Yongqi Yuan ,&nbsp;Ge Jin ,&nbsp;Huan Wang","doi":"10.1016/j.compag.2026.111431","DOIUrl":"10.1016/j.compag.2026.111431","url":null,"abstract":"<div><div>Effective feature representation and heterogeneous fusion are essential for plant leaf recognition. However, existing methods have several limitations, such as insufficient comprehensiveness and distinctiveness in feature representation, as well as a lack of full consideration for the compatibility and complementarity in heterogeneous fusion. In the end, we propose a discriminative shape representation named the bag of multiscale curvature angle cuts (BMCAC) to capture fine curvature and spatial distribution characteristics, an advanced deep representation called the progressive salient deep representation (PSDR) to fully exploit deep convolutional features, and an effective fusion framework termed the K-weighted shape and deep feature fusion (KWFF) to aggregate the local context and global importance of heterogeneous features. Specifically, BMCAC is derived from the curvature angle cuts (CAC), multiscale analysis, and the bag of visual words (BoVW) model; PSDR is constructed by applying progressive downsampling and hierarchical pooling operations to deep convolutional features; and KWFF is developed by encoding neighboring information using homogeneous distance measures while incorporating globally weighted contributions from heterogeneous distance measures. Extensive experiments on four well-known benchmark leaf datasets demonstrate that the proposed shape and deep representations can efficiently extract leaf image features, and the fusion framework can effectively integrate heterogeneous features, outperforming state-of-the-art methods. The source code is available at <span><span>https://github.com/Mumuxi1123/BMCAC_PSDR_KWFF</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":"244 ","pages":"Article 111431"},"PeriodicalIF":8.9,"publicationDate":"2026-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146025083","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computers and Electronics in Agriculture
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1