首页 > 最新文献

Computers and Electronics in Agriculture最新文献

英文 中文
Real-time lettuce-weed localization and weed severity classification based on lightweight YOLO convolutional neural networks for intelligent intra-row weed control 基于轻量级 YOLO 卷积神经网络的生菜杂草实时定位和杂草严重程度分类,实现智能行内杂草控制
IF 7.7 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2024-09-04 DOI: 10.1016/j.compag.2024.109404

Competition for nutrients between intra-row weeds and cultivated vegetables is a major contributor to reduced crop yields. Compared with manual weeding, intelligent robots can improve the efficiency of weeding operations. Developing real-time and reliable robotic systems for weed control in vegetable fields is a significant challenge due to the complexity of real-time identification, localization, and classification of vegetables as well as various weed species. The main purpose of this study was to propose a high-performance, lightweight deep learning model and an intra-row weed severity classification algorithm for real-time lettuce identification and weed severity classification. In this study, a scaling factor (τ = 0.5) was chosen to lightweight the YOLOv7 model. A new Multimodule-YOLOv7-L lightweight model was then developed by combining ECA and CA attention mechanisms, ELAN-B3 and DownC modules. The overall performance of the Multimodule-YOLOv7- L was better than that of other deep learning models, including YOLOv7, YOLOv7-Tiny, YOLOv8m, YOLOv5n-Cabbage, SE: YOLOv5x, YOLOv5s_Ghb, MST-YOLO_CBAM, Citrus-YOLOv7, Pineapple-YOLOv7, MS-YOLOv7 and CBAM-YOLOv7. The precision, recall, [email protected], F1-score, model weight and FPS of the Multimodule-YOLOv7- L model were 97.5 %, 95.7 %, 97.1 %, 96.6 %, 18.4 MB and 37.3 FPS (Image resolution about 3000 × 3000), respectively. An intra-row weed severity classification algorithm based on the Multimodule-YOLOv7-L model was proposed for use in a new mechanical-laser collaborative intra-row weeding robot. The developed algorithm achieved a classification accuracy of 100 % in eight lettuce weed scenarios, with the processing time of a single image ranging from 4-13 ms. The results of this study provided valuable reference for the development of intelligent robots for intra-row weed control. The algorithm proposed in this article can be obtained at https://github.com/H777R/The-intra-row-weed-severity-classification-algorithm.git.

行内杂草与栽培蔬菜争夺养分是导致作物减产的主要原因。与人工除草相比,智能机器人可以提高除草作业的效率。由于实时识别、定位和分类蔬菜以及各种杂草的复杂性,开发实时可靠的机器人系统来控制菜田杂草是一项重大挑战。本研究的主要目的是提出一种高性能、轻量级的深度学习模型和行内杂草严重程度分类算法,用于实时识别生菜和进行杂草严重程度分类。本研究选择了一个缩放因子(τ = 0.5)来实现 YOLOv7 模型的轻量化。然后,结合 ECA 和 CA 注意机制、ELAN-B3 和 DownC 模块,开发了一种新的多模块 YOLOv7-L 轻量级模型。多模块-YOLOv7-L的整体性能优于其他深度学习模型,包括YOLOv7、YOLOv7-Tiny、YOLOv8m、YOLOv5n-Cabbage、SE: YOLOv5x、YOLOv5s_Ghb、MST-YOLO_CBAM、Citrus-YOLOv7、Pineapple-YOLOv7、MS-YOLOv7和CBAM-YOLOv7。多模块-YOLOv7- L 模型的精确度、召回率、[email protected]、F1 分数、模型权重和 FPS 分别为 97.5%、95.7%、97.1%、96.6%、18.4 MB 和 37.3 FPS(图像分辨率约为 3000 × 3000)。提出了一种基于多模块-YOLOv7-L 模型的行内杂草严重程度分类算法,用于新型机械激光协作行内除草机器人。所开发的算法在八种生菜杂草情况下的分类准确率达到 100%,单张图像的处理时间为 4-13 毫秒。该研究结果为开发用于行内除草的智能机器人提供了宝贵的参考。本文提出的算法可在 https://github.com/H777R/The-intra-row-weed-severity-classification-algorithm.git 上获取。
{"title":"Real-time lettuce-weed localization and weed severity classification based on lightweight YOLO convolutional neural networks for intelligent intra-row weed control","authors":"","doi":"10.1016/j.compag.2024.109404","DOIUrl":"10.1016/j.compag.2024.109404","url":null,"abstract":"<div><p>Competition for nutrients between intra-row weeds and cultivated vegetables is a major contributor to reduced crop yields. Compared with manual weeding, intelligent robots can improve the efficiency of weeding operations. Developing real-time and reliable robotic systems for weed control in vegetable fields is a significant challenge due to the complexity of real-time identification, localization, and classification of vegetables as well as various weed species. The main purpose of this study was to propose a high-performance, lightweight deep learning model and an intra-row weed severity classification algorithm for real-time lettuce identification and weed severity classification. In this study, a scaling factor (<em>τ</em> = 0.5) was chosen to lightweight the YOLOv7 model. A new Multimodule-YOLOv7-L lightweight model was then developed by combining ECA and CA attention mechanisms, ELAN-B3 and DownC modules. The overall performance of the Multimodule-YOLOv7- L was better than that of other deep learning models, including YOLOv7, YOLOv7-Tiny, YOLOv8m, YOLOv5n-Cabbage, SE: YOLOv5x, YOLOv5s_Ghb, MST-YOLO_CBAM, Citrus-YOLOv7, Pineapple-YOLOv7, MS-YOLOv7 and CBAM-YOLOv7. The precision, recall, [email protected], F1-score, model weight and FPS of the Multimodule-YOLOv7- L model were 97.5 %, 95.7 %, 97.1 %, 96.6 %, 18.4 MB and 37.3 FPS (Image resolution about 3000 × 3000), respectively. An intra-row weed severity classification algorithm based on the Multimodule-YOLOv7-L model was proposed for use in a new mechanical-laser collaborative intra-row weeding robot. The developed algorithm achieved a classification accuracy of 100 % in eight lettuce weed scenarios, with the processing time of a single image ranging from 4-13 ms. The results of this study provided valuable reference for the development of intelligent robots for intra-row weed control. The algorithm proposed in this article can be obtained at <span><span>https://github.com/H777R/The-intra-row-weed-severity-classification-algorithm.git</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":null,"pages":null},"PeriodicalIF":7.7,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142137251","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Three-dimensional image analysis for almond endocarp feature extraction and shape description 用于杏仁内果皮特征提取和形状描述的三维图像分析
IF 7.7 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2024-09-04 DOI: 10.1016/j.compag.2024.109420

We propose a morphological characterization of the endocarp of the fruit of the almond tree, Prunus amygdalus (Batsch), using computer vision techniques to extract features in 3D almond endocarp meshes with the objective to describe the diversity of the crop in a systematic and unambiguous form. All the proposed descriptors are quantitative and easily computable, allowing fast and objective assessments of the morphological variations between almond varieties. We collect and 3D-scan a total of 9510 almond endocarps to obtain such meshes, to which we apply an affine transformation so that they are positioned in a standardized reference where meaningful physical measures can be taken. Complex descriptors derived from the geometry of the endocarp are then introduced to identify richer features. The use of 3D, compared to simply taking 2D images, allows for a more accurate and complete description of the endocarp shape. In particular, the contour and apex shapes, keel development, markings on the surface, and symmetry of the endocarp are analyzed and given quantitative measures. The validity of the presented morphological descriptors is finally tested on 2610 endocarps from the collected dataset, corresponding to 36 autochthonous almond varieties from the island of Mallorca (Spain) and 14 international reference varieties, all with well documented characteristics. Numerical results show that the proposed descriptors agree with human-made shape classifications of the studied varieties with a coincidence of 75.0% for contour shape, 76.0% for apex shape, and 80.0% for keel development. Visual comparisons of the extracted features also show that they are coherent with commonly used guidelines for the morphological characterization of the almond endocarp. We conclude that the use of 3D imaging approaches for the description of the almond endocarp is a promising alternative to traditional methods, providing a reliable way to deal with ambiguity and helping reduce biases and inconsistencies caused by subjective visual evaluations.

我们提出了一种杏树(Prunus amygdalus (Batsch))果实内果皮形态特征描述方法,利用计算机视觉技术提取三维杏树内果皮网格中的特征,目的是以系统和明确的形式描述作物的多样性。所有建议的描述符都是定量的,易于计算,可以快速客观地评估杏仁品种间的形态变化。我们共收集并三维扫描了 9510 个杏仁内果皮,获得了这些网格,并对其进行了仿射变换,从而将其定位在一个标准化的参照物上,以便进行有意义的物理测量。然后引入从内果皮几何形状中提取的复杂描述符,以识别更丰富的特征。与简单拍摄二维图像相比,使用三维图像可以更准确、更完整地描述果皮的形状。特别是对内果皮的轮廓和顶点形状、龙骨发育、表面标记和对称性进行了分析和定量测量。最后,对收集的数据集中的 2610 个内果皮进行了测试,这些内果皮对应于马略卡岛(西班牙)的 36 个本地杏仁品种和 14 个国际参考品种,所有这些品种的特征都有据可查。数值结果显示,所提出的描述符与所研究品种的人工形状分类一致,轮廓形状吻合度为 75.0%,顶点形状吻合度为 76.0%,龙骨发育吻合度为 80.0%。对提取的特征进行目视比较也表明,它们与常用的杏仁内果皮形态特征描述指南一致。我们得出的结论是,使用三维成像方法描述杏仁内果皮是一种替代传统方法的有前途的方法,它提供了一种处理模糊性的可靠方法,并有助于减少主观视觉评价造成的偏差和不一致。
{"title":"Three-dimensional image analysis for almond endocarp feature extraction and shape description","authors":"","doi":"10.1016/j.compag.2024.109420","DOIUrl":"10.1016/j.compag.2024.109420","url":null,"abstract":"<div><p>We propose a morphological characterization of the endocarp of the fruit of the almond tree, <em>Prunus amygdalus</em> (Batsch), using computer vision techniques to extract features in 3D almond endocarp meshes with the objective to describe the diversity of the crop in a systematic and unambiguous form. All the proposed descriptors are quantitative and easily computable, allowing fast and objective assessments of the morphological variations between almond varieties. We collect and 3D-scan a total of 9510 almond endocarps to obtain such meshes, to which we apply an affine transformation so that they are positioned in a standardized reference where meaningful physical measures can be taken. Complex descriptors derived from the geometry of the endocarp are then introduced to identify richer features. The use of 3D, compared to simply taking 2D images, allows for a more accurate and complete description of the endocarp shape. In particular, the contour and apex shapes, keel development, markings on the surface, and symmetry of the endocarp are analyzed and given quantitative measures. The validity of the presented morphological descriptors is finally tested on 2610 endocarps from the collected dataset, corresponding to 36 autochthonous almond varieties from the island of Mallorca (Spain) and 14 international reference varieties, all with well documented characteristics. Numerical results show that the proposed descriptors agree with human-made shape classifications of the studied varieties with a coincidence of <span><math><mrow><mtext>75.0</mtext><mspace></mspace><mtext>%</mtext></mrow></math></span> for contour shape, <span><math><mrow><mtext>76.0</mtext><mspace></mspace><mtext>%</mtext></mrow></math></span> for apex shape, and <span><math><mrow><mtext>80.0</mtext><mspace></mspace><mtext>%</mtext></mrow></math></span> for keel development. Visual comparisons of the extracted features also show that they are coherent with commonly used guidelines for the morphological characterization of the almond endocarp. We conclude that the use of 3D imaging approaches for the description of the almond endocarp is a promising alternative to traditional methods, providing a reliable way to deal with ambiguity and helping reduce biases and inconsistencies caused by subjective visual evaluations.</p></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":null,"pages":null},"PeriodicalIF":7.7,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0168169924008111/pdfft?md5=fc6e130c431b7d7cd003bd9f0b88c9b4&pid=1-s2.0-S0168169924008111-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142137254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FCOS-EAM: An accurate segmentation method for overlapping green fruits FCOS-EAM:重叠绿色水果的精确分割方法
IF 7.7 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2024-09-04 DOI: 10.1016/j.compag.2024.109392

Accurate fruit detection and segmentation based on deep learning is the key to successful harvesting robot operations, but the complex background of orchards, light and branch shading, and fruit overlap lead to low detection and segmentation accuracy and high complexity of existing methods. To address these problems, an improved green fruit segmentation method based on FCOS is proposed in this study. Firstly, its segmentation function for green fruits is realized by adding segmentation module. Then, the FCOS head network is improved by adding the Border-attention module (BAM) to detect the boundary of green fruits with higher accuracy. In addition, the features of mask branch and edge segmentation branch are fused in the segmentation module, and the appearance commonality is learned by modeling the pairwise affinity between all pixels of the feature map using non-local affinity-parsing, and finally the segmentation prediction results are output by combining the feature map of fruit shape and appearance commonality. The experimental results show that this model achieves 81.2% segmentation accuracy on apple dataset and 77.9% segmentation accuracy on persimmon dataset with relatively low guarantee complexity, which exceeds most current segmentation models. Meanwhile, this model has high robustness and can be used for the detection and segmentation work of other green fruits and vegetables in orchards, while effectively extending the application of agricultural equipment.

基于深度学习的精确水果检测和分割是收获机器人成功作业的关键,但果园背景复杂、光照和树枝遮挡、水果重叠等因素导致现有方法的检测和分割精度低、复杂度高。针对这些问题,本研究提出了一种基于 FCOS 的改进型绿色水果分割方法。首先,通过添加分割模块实现其对绿色水果的分割功能。然后,通过添加边界注意模块(BAM)对 FCOS 头网络进行改进,以更高的精度检测绿色水果的边界。此外,在分割模块中还融合了掩膜分支和边缘分割分支的特征,并利用非局部亲和性解析对特征图中所有像素点之间的成对亲和性进行建模,从而学习外观共性,最后结合水果形状特征图和外观共性输出分割预测结果。实验结果表明,该模型在苹果数据集上的分割准确率达到 81.2%,在柿子数据集上的分割准确率达到 77.9%,保证复杂度相对较低,超过了目前大多数分割模型。同时,该模型具有较高的鲁棒性,可用于果园中其他绿色果蔬的检测和分割工作,同时有效扩展了农业设备的应用范围。
{"title":"FCOS-EAM: An accurate segmentation method for overlapping green fruits","authors":"","doi":"10.1016/j.compag.2024.109392","DOIUrl":"10.1016/j.compag.2024.109392","url":null,"abstract":"<div><p>Accurate fruit detection and segmentation based on deep learning is the key to successful harvesting robot operations, but the complex background of orchards, light and branch shading, and fruit overlap lead to low detection and segmentation accuracy and high complexity of existing methods. To address these problems, an improved green fruit segmentation method based on FCOS is proposed in this study. Firstly, its segmentation function for green fruits is realized by adding segmentation module. Then, the FCOS head network is improved by adding the Border-attention module (BAM) to detect the boundary of green fruits with higher accuracy. In addition, the features of mask branch and edge segmentation branch are fused in the segmentation module, and the appearance commonality is learned by modeling the pairwise affinity between all pixels of the feature map using non-local affinity-parsing, and finally the segmentation prediction results are output by combining the feature map of fruit shape and appearance commonality. The experimental results show that this model achieves 81.2% segmentation accuracy on apple dataset and 77.9% segmentation accuracy on persimmon dataset with relatively low guarantee complexity, which exceeds most current segmentation models. Meanwhile, this model has high robustness and can be used for the detection and segmentation work of other green fruits and vegetables in orchards, while effectively extending the application of agricultural equipment.</p></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":null,"pages":null},"PeriodicalIF":7.7,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142137252","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predictions of apple mechanical damage volume using micro-CT measurements and support vector regression(SVR) 利用微型计算机断层扫描测量和支持向量回归(SVR)预测苹果机械损伤体积
IF 7.7 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2024-09-03 DOI: 10.1016/j.compag.2024.109402

Accurately calculating the damage volume and making clear the interconnected effects of the physical and chemical properties of apples on mechanical damage are crucial steps in reducing the possibility of apple damage. Tests have been conducted on apples at different maturity levels, including measuring the firmness, moisture content, water-soluble pectin (WSP) content, soluble solids content (SSC) of the flesh, and elastic modulus of the apple flesh and peel. Transient collisions were performed using a pendulum device to create damage zones under specific impact energies. Then, the X-ray micro-computed tomography (Micro-CT) was utilized to quantitatively analyse mechanical damage volumes, the effects of apple tissue characteristics and impact energy on the damage volume were analysed in detail. The results indicated that higher-maturity apples were more susceptible to mechanical damage, and Micro-CT measurements were more accurate when the impact energy ≥ 0.05 J, while the empirical formula showed greater deviation; the curvature radius at the impact point can be considered as a latent variable influencing the apple damage volume. Furthermore, a damage volume prediction model, based on bruise area calculated by the empirical formula, WSP content of the flesh, and elastic modulus of the apple flesh and peel, was established. With a testing dataset without anticipate in model training for verification, the developed model achieved a coefficient of determination of 0.9782, indicating that the model can predict damage volume effectively and reduce errors associated with the empirical formula, particularly at higher impact energies. The research can provide insights into potential applications in apple industry practices to reduce the mechanical damage.

准确计算损伤量并明确苹果的物理和化学特性对机械损伤的相互影响,是减少苹果损伤可能性的关键步骤。已对不同成熟度的苹果进行了测试,包括测量苹果果肉和果皮的硬度、水分含量、水溶性果胶 (WSP) 含量、果肉可溶性固形物含量 (SSC) 以及弹性模量。使用摆锤装置进行瞬态碰撞,在特定的冲击能量下产生损伤区。然后,利用 X 射线显微计算机断层扫描(Micro-CT)对机械损伤体积进行定量分析,并详细分析了苹果组织特征和冲击能量对损伤体积的影响。结果表明,成熟度较高的苹果更容易受到机械损伤,当冲击能量≥0.05 J时,Micro-CT测量结果更准确,而经验公式则出现较大偏差;冲击点的曲率半径可视为影响苹果损伤体积的潜在变量。此外,根据经验公式计算出的碰伤面积、果肉中的 WSP 含量以及苹果果肉和果皮的弹性模量,建立了损伤体积预测模型。通过对模型训练中没有预期的测试数据集进行验证,所建立的模型达到了 0.9782 的决定系数,表明该模型可以有效地预测损伤体积,减少与经验公式相关的误差,尤其是在冲击能量较高的情况下。这项研究可为苹果产业实践中减少机械损伤的潜在应用提供启示。
{"title":"Predictions of apple mechanical damage volume using micro-CT measurements and support vector regression(SVR)","authors":"","doi":"10.1016/j.compag.2024.109402","DOIUrl":"10.1016/j.compag.2024.109402","url":null,"abstract":"<div><p>Accurately calculating the damage volume and making clear the interconnected effects of the physical and chemical properties of apples on mechanical damage are crucial steps in reducing the possibility of apple damage. Tests have been conducted on apples at different maturity levels, including measuring the firmness, moisture content, water-soluble pectin (WSP) content, soluble solids content (SSC) of the flesh, and elastic modulus of the apple flesh and peel. Transient collisions were performed using a pendulum device to create damage zones under specific impact energies. Then, the X-ray micro-computed tomography (Micro-CT) was utilized to quantitatively analyse mechanical damage volumes, the effects of apple tissue characteristics and impact energy on the damage volume were analysed in detail. The results indicated that higher-maturity apples were more susceptible to mechanical damage, and Micro-CT measurements were more accurate when the impact energy ≥ 0.05 J, while the empirical formula showed greater deviation; the curvature radius at the impact point can be considered as a latent variable influencing the apple damage volume. Furthermore, a damage volume prediction model, based on bruise area calculated by the empirical formula, WSP content of the flesh, and elastic modulus of the apple flesh and peel, was established. With a testing dataset without anticipate in model training for verification, the developed model achieved a coefficient of determination of 0.9782, indicating that the model can predict damage volume effectively and reduce errors associated with the empirical formula, particularly at higher impact energies. The research can provide insights into potential applications in apple industry practices to reduce the mechanical damage.</p></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":null,"pages":null},"PeriodicalIF":7.7,"publicationDate":"2024-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142129978","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spiking-LSTM: A novel hyperspectral image segmentation network for Sclerotinia detection Spiking-LSTM:用于检测硬菌的新型高光谱图像分割网络
IF 7.7 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2024-09-03 DOI: 10.1016/j.compag.2024.109397

Sclerotinia is a worldwide disease that often occurs at all growth stages of rapeseed, and can lead to 10 %∼70 % yield decline. It will also drastically reduce the oil content of seeds, which greatly increases the risk and difficulty of rapeseed cultivation. In order to address the problems of traditional chemical-based Sclerotinia detection methods such as complex operation, environmental pollution, plant damage and low efficiency, this study innovatively combined the two architectures of SNN (Spiking Neural Network) and LSTM, and proposed a spatial-spectral joint detection model Spiking-LSTM for the HSI segmentation. In the design process of the model, spiking neurons were used instead of gating functions in traditional LSTM units, while the back propagation of errors was solved by using gradient surrogate function. The experimental results show that when input data from the infected area, the neurons in hidden layer of the trained model exhibited distinctly regular spiking signals. Compared with the mainstream models, the Spiking-LSTM based on spatial-spectral data fusion has better performance in the evaluation parameters such as mAP, ClassAP, mIoU, FWIoU and Kappa coefficient. Its Sclerotinia detection mAP reached 94.3 % and was able to accurately extract the infected areas at the early-stage of infection. With essentially the same structure, the Spiking LSTM not only has higher detection accuracy but also, for the same HSI input, requires only one-fifth of the theoretical energy consumption compared to the traditional LSTM. This paper establishes the basis for the construction of large-scale SNN models, and also provides a reference for the application of SNNs in different fields.

硬核病是一种世界性病害,通常发生在油菜籽的各个生长阶段,可导致产量下降 10%∼70%。它还会大幅降低种子的含油量,大大增加了油菜籽种植的风险和难度。针对传统基于化学方法的赤霉病检测方法操作复杂、污染环境、损害植株、检测效率低等问题,本研究创新性地结合了SNN(尖峰神经网络)和LSTM两种体系结构,提出了一种用于HSI分割的空间-光谱联合检测模型Spiking-LSTM。在该模型的设计过程中,使用了尖峰神经元来代替传统 LSTM 单元中的门控函数,同时使用梯度代理函数来解决误差反向传播问题。实验结果表明,当输入来自疫区的数据时,训练好的模型隐层神经元表现出明显规则的尖峰信号。与主流模型相比,基于空间-光谱数据融合的 Spiking-LSTM 在 mAP、ClassAP、mIoU、FWIoU 和 Kappa 系数等评价参数上都有更好的表现。其硬菌检测 mAP 达到 94.3%,能够在感染初期准确提取感染区域。在结构基本相同的情况下,Spiking LSTM 不仅具有更高的检测精度,而且在相同的 HSI 输入条件下,其理论能耗仅为传统 LSTM 的五分之一。本文为构建大规模 SNN 模型奠定了基础,也为 SNN 在不同领域的应用提供了参考。
{"title":"Spiking-LSTM: A novel hyperspectral image segmentation network for Sclerotinia detection","authors":"","doi":"10.1016/j.compag.2024.109397","DOIUrl":"10.1016/j.compag.2024.109397","url":null,"abstract":"<div><p>Sclerotinia is a worldwide disease that often occurs at all growth stages of rapeseed, and can lead to 10 %∼70 % yield decline. It will also drastically reduce the oil content of seeds, which greatly increases the risk and difficulty of rapeseed cultivation. In order to address the problems of traditional chemical-based Sclerotinia detection methods such as complex operation, environmental pollution, plant damage and low efficiency, this study innovatively combined the two architectures of SNN (Spiking Neural Network) and LSTM, and proposed a spatial-spectral joint detection model Spiking-LSTM for the HSI segmentation. In the design process of the model, spiking neurons were used instead of gating functions in traditional LSTM units, while the back propagation of errors was solved by using gradient surrogate function. The experimental results show that when input data from the infected area, the neurons in hidden layer of the trained model exhibited distinctly regular spiking signals. Compared with the mainstream models, the Spiking-LSTM based on spatial-spectral data fusion has better performance in the evaluation parameters such as mAP, ClassAP, mIoU, FWIoU and Kappa coefficient. Its Sclerotinia detection mAP reached 94.3 % and was able to accurately extract the infected areas at the early-stage of infection. With essentially the same structure, the Spiking LSTM not only has higher detection accuracy but also, for the same HSI input, requires only one-fifth of the theoretical energy consumption compared to the traditional LSTM. This paper establishes the basis for the construction of large-scale SNN models, and also provides a reference for the application of SNNs in different fields.</p></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":null,"pages":null},"PeriodicalIF":7.7,"publicationDate":"2024-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142129967","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MAFDE-DN4: Improved Few-shot plant disease classification method based on Deep Nearest Neighbor Neural Network MAFDE-DN4:基于深度近邻神经网络的改进型多发植物病害分类方法
IF 7.7 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2024-09-02 DOI: 10.1016/j.compag.2024.109373

Deep learning-based methods for accurately identifying plant diseases can be effective in improving crop yields. However, the effectiveness of these methods heavily relies on the availability of large-scale manually labeled datasets, which present technical and economic challenges. Few-shot learning methods can be generalized to new categories with a small number of samples, which is very promising in the field of plant disease recognition despite the limited sample size. However, due to the complexity of real-life scenarios, distribution of disease leaves, and significant intra-class variability and inter-class similarity resulting from crop species and disease species, existing methods perform poorly in the field of plant disease recognition. To address the above problems, this paper proposes an improved multi-scale attention fusion with discriminative enhancement deep nearest neighbor neural network MAFDE-DN4 based on DN4. Our approach makes three main contributions. First, we have designed a bidirectional weighted feature fusion module (BWFM) to enhance the aggregation of fine-grained features and enhance the network’s representation of complex disease images. Second, to tackle the issue of sparse feature descriptors being vulnerable to irrelevant noise in small sample conditions, a episodic attention module (EA) has been developed to produce scene category-relevant attention maps. This effectively mitigates the influence of irrelevant background information. Finally, we introduce additional spacing between category margins to enhance the original softmax loss function, amplify the inter-class differences to reduce the intra-class distances, and add L2 regularization constraint terms to stabilize the training process. To simulate different real-world scenarios, we set up different dataset settings. Under the 1-shot task and the 5-shot task, our method achieves 57.5% and 81.41% accuracy under the within-domain strategy and 36.54% and 51.23% accuracy under the cross-domain strategy. The experimental results show that our method outperforms existing related work in the field of plant disease recognition, whether it is a dataset with a single background or a field dataset with a complex background in a real scenario. MAFDE-DN4 based on Few-shot learning requires substantially less data on new categories of plant diseases.

基于深度学习的植物病害准确识别方法可以有效提高作物产量。然而,这些方法的有效性在很大程度上依赖于大规模人工标注数据集的可用性,这带来了技术和经济上的挑战。少量学习方法只需少量样本就能归纳出新的类别,尽管样本量有限,但在植物病害识别领域却大有可为。然而,由于现实场景的复杂性、病叶的分布以及作物种类和病害种类导致的显著类内变异性和类间相似性,现有方法在植物病害识别领域表现不佳。针对上述问题,本文提出了一种基于 DN4 的改进型多尺度注意力融合与判别增强深度近邻神经网络 MAFDE-DN4。我们的方法主要有三个贡献。首先,我们设计了一个双向加权特征融合模块(BWFM),以增强细粒度特征的聚合,提高网络对复杂疾病图像的表示能力。其次,为了解决稀疏特征描述符在小样本条件下易受无关噪声影响的问题,我们开发了一个表观注意力模块(EA),以生成与场景类别相关的注意力图。这有效地减轻了无关背景信息的影响。最后,我们引入了类别边缘之间的额外间距,以增强原始的 softmax 损失函数,放大类间差异以缩小类内距离,并添加 L2 正则化约束项以稳定训练过程。为了模拟现实世界的不同场景,我们设置了不同的数据集。在一枪任务和五枪任务中,我们的方法在域内策略下分别取得了 57.5% 和 81.41% 的准确率,在跨域策略下分别取得了 36.54% 和 51.23% 的准确率。实验结果表明,在植物病害识别领域,无论是单一背景的数据集,还是实际场景中复杂背景的田间数据集,我们的方法都优于现有的相关工作。基于 Few-shot 学习的 MAFDE-DN4 对植物病害新类别所需的数据量大大减少。
{"title":"MAFDE-DN4: Improved Few-shot plant disease classification method based on Deep Nearest Neighbor Neural Network","authors":"","doi":"10.1016/j.compag.2024.109373","DOIUrl":"10.1016/j.compag.2024.109373","url":null,"abstract":"<div><p>Deep learning-based methods for accurately identifying plant diseases can be effective in improving crop yields. However, the effectiveness of these methods heavily relies on the availability of large-scale manually labeled datasets, which present technical and economic challenges. Few-shot learning methods can be generalized to new categories with a small number of samples, which is very promising in the field of plant disease recognition despite the limited sample size. However, due to the complexity of real-life scenarios, distribution of disease leaves, and significant intra-class variability and inter-class similarity resulting from crop species and disease species, existing methods perform poorly in the field of plant disease recognition. To address the above problems, this paper proposes an improved multi-scale attention fusion with discriminative enhancement deep nearest neighbor neural network MAFDE-DN4 based on DN4. Our approach makes three main contributions. First, we have designed a bidirectional weighted feature fusion module (BWFM) to enhance the aggregation of fine-grained features and enhance the network’s representation of complex disease images. Second, to tackle the issue of sparse feature descriptors being vulnerable to irrelevant noise in small sample conditions, a episodic attention module (EA) has been developed to produce scene category-relevant attention maps. This effectively mitigates the influence of irrelevant background information. Finally, we introduce additional spacing between category margins to enhance the original softmax loss function, amplify the inter-class differences to reduce the intra-class distances, and add L2 regularization constraint terms to stabilize the training process. To simulate different real-world scenarios, we set up different dataset settings. Under the 1-shot task and the 5-shot task, our method achieves 57.5% and 81.41% accuracy under the within-domain strategy and 36.54% and 51.23% accuracy under the cross-domain strategy. The experimental results show that our method outperforms existing related work in the field of plant disease recognition, whether it is a dataset with a single background or a field dataset with a complex background in a real scenario. MAFDE-DN4 based on Few-shot learning requires substantially less data on new categories of plant diseases.</p></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":null,"pages":null},"PeriodicalIF":7.7,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142122003","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Lightweight cabbage segmentation network and improved weed detection method 轻量级卷心菜分割网络和改进的杂草检测方法
IF 7.7 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2024-09-02 DOI: 10.1016/j.compag.2024.109403

This study addressed the challenge of machine vision-based weed detection for precision herbicide application, a task complicated by the diversity of weed species, ecotypes, and variations in growth stages. We propose an indirect approach that segments crops and classifies the remaining green objects as weeds. A novel, lightweight segmentation network was developed to reduce computational demands without compromising accuracy. The model, with a size of just 2.64 MB, achieves an impressive mean Intersection over Union (mIoU) of 97.9 %, with a recall of 93.4 %, and a precision of 97.6 %, while also enhancing inference speed. Subsequently, improvements were implemented using the image processing method for extracting green plants. A crop mask was generated using a segmentation algorithm, and a mask expansion mechanism was introduced to rectify errors in the initial phase of crop segmentation. A cost-effective threshold adjustment operation was applied to eliminate the environmental influences on the detection results. The results indicate that the weed detection method completely avoided the complexity related to the variations in species, ecotypes, growth stages, and densities of weeds across different fields and realized accurate, effective, and reliable weed detection in cabbage.

这项研究旨在解决基于机器视觉的杂草检测难题,以实现除草剂的精准施用,而杂草种类、生态型和生长阶段的多样性使这项任务变得更加复杂。我们提出了一种间接方法,即分割农作物并将剩余的绿色物体归类为杂草。我们开发了一种新颖的轻量级分割网络,在不影响准确性的前提下降低了计算需求。该模型的大小仅为 2.64 MB,却达到了令人印象深刻的平均交集大于联合率 (mIoU) 97.9%,召回率 93.4%,精确率 97.6%,同时还提高了推理速度。随后,使用提取绿色植物的图像处理方法进行了改进。使用分割算法生成作物掩膜,并引入掩膜扩展机制来纠正作物分割初始阶段的错误。为了消除环境对检测结果的影响,采用了一种经济有效的阈值调整操作。结果表明,该杂草检测方法完全避免了不同田间杂草的种类、生态型、生长阶段和密度差异所带来的复杂性,实现了准确、有效和可靠的白菜杂草检测。
{"title":"Lightweight cabbage segmentation network and improved weed detection method","authors":"","doi":"10.1016/j.compag.2024.109403","DOIUrl":"10.1016/j.compag.2024.109403","url":null,"abstract":"<div><p>This study addressed the challenge of machine vision-based weed detection for precision herbicide application, a task complicated by the diversity of weed species, ecotypes, and variations in growth stages. We propose an indirect approach that segments crops and classifies the remaining green objects as weeds. A novel, lightweight segmentation network was developed to reduce computational demands without compromising accuracy. The model, with a size of just 2.64 MB, achieves an impressive mean Intersection over Union (mIoU) of 97.9 %, with a recall of 93.4 %, and a precision of 97.6 %, while also enhancing inference speed. Subsequently, improvements were implemented using the image processing method for extracting green plants. A crop mask was generated using a segmentation algorithm, and a mask expansion mechanism was introduced to rectify errors in the initial phase of crop segmentation. A cost-effective threshold adjustment operation was applied to eliminate the environmental influences on the detection results. The results indicate that the weed detection method completely avoided the complexity related to the variations in species, ecotypes, growth stages, and densities of weeds across different fields and realized accurate, effective, and reliable weed detection in cabbage.</p></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":null,"pages":null},"PeriodicalIF":7.7,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142122002","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unsupervised anomaly detection for pome fruit quality inspection using X-ray radiography 利用 X 射线射线照相术进行无监督异常检测,以检测果核类水果的质量
IF 7.7 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2024-09-02 DOI: 10.1016/j.compag.2024.109364

A novel fully convolutional autoencoder (convAE) was introduced to analyze X-ray radiography images of ‘Braeburn’ apples and ‘Conference’ pears with and without disorders for online sorting purposes. The model was solely trained on either apple or pear samples without disorders and outperformed a traditional autoencoder (AE) across multiple test sets. We evaluated our approach using the area under the curve (AUC) as an evaluation metric. A cross-test experiment further demonstrated consistent performance between a model trained on apple data for classifying pear fruit (accuracy: 71 %) and a pear-specific model (accuracy: 70 %). We also evaluated models trained on simulated X-ray radiographs with real ones, and vice versa. For instance, under scenario of training on real data and testing on simulated X-ray radiographs, an accuracy of 80 % for detecting disordered non-consumable pear was achieved. This work provides valuable insights into anomaly detection for apples and pears with several disorders.

该研究引入了一种新型全卷积自动编码器(convAE),用于分析有无病变的 "Braeburn "苹果和 "Conference "梨的 X 射线射线图像,以实现在线分拣的目的。该模型仅在无病变的苹果或梨样本上进行了训练,在多个测试集上的表现优于传统的自动编码器 (AE)。我们使用曲线下面积(AUC)作为评估指标对我们的方法进行了评估。交叉测试实验进一步证明,在苹果数据上训练的梨果分类模型(准确率:71%)和梨果专用模型(准确率:70%)的性能一致。我们还用真实的 X 射线照片评估了在模拟 X 射线照片上训练的模型,反之亦然。例如,在使用真实数据进行训练和使用模拟 X 射线照片进行测试的情况下,检测无序非食用梨的准确率达到了 80%。这项工作为苹果和梨的异常检测提供了宝贵的见解。
{"title":"Unsupervised anomaly detection for pome fruit quality inspection using X-ray radiography","authors":"","doi":"10.1016/j.compag.2024.109364","DOIUrl":"10.1016/j.compag.2024.109364","url":null,"abstract":"<div><p>A novel fully convolutional autoencoder (convAE) was introduced to analyze X-ray radiography images of ‘Braeburn’ apples and ‘Conference’ pears with and without disorders for online sorting purposes. The model was solely trained on either apple or pear samples without disorders and outperformed a traditional autoencoder (AE) across multiple test sets. We evaluated our approach using the a<em>rea under the curve</em> (AUC) as an evaluation metric. A cross-test experiment further demonstrated consistent performance between a model trained on apple data for classifying pear fruit (accuracy: 71 %) and a pear-specific model (accuracy: 70 %). We also evaluated models trained on simulated X-ray radiographs with real ones, and vice versa. For instance, under scenario of training on real data and testing on simulated X-ray radiographs, an accuracy of 80 % for detecting disordered non-consumable pear was achieved. This work provides valuable insights into anomaly detection for apples and pears with several disorders.</p></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":null,"pages":null},"PeriodicalIF":7.7,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142129966","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Motion focus global–local network: Combining attention mechanism with micro action features for cow behavior recognition 运动聚焦全局-局部网络:结合注意力机制和微动作特征识别奶牛行为
IF 7.7 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2024-09-01 DOI: 10.1016/j.compag.2024.109399

The use of machine vision technology for recognizing cow behavior plays a crucial role in daily management, health monitoring, and breeding and reproduction in dairy farming, making it an essential component of modern smart agriculture. This paper presents a novel dual-stream network model, the Motion Focus Global-Local Network (MFGN), for analyzing cow video data. The dual-stream network consists of a global spatiotemporal feature stream and a fine motion feature stream. The global spatiotemporal feature stream extracts key frames to remove redundant information and utilizes a Transformer network for global spatio-temporal feature extraction, reflecting the dynamic changes in cow videos and the temporal relationships between video frames. The fine motion feature stream, based on frame differencing of cow videos, uses focal convolution to capture subtle movements of cows, enhancing the focus on minor behavioral changes. To evaluate the performance of the proposed model, video data samples were collected from eight cows marked on their bodies and heads at an Australian farm site (CSIRO Armidale), including a total of 1715 video sequences across three behavior categories. The model achieved recognition accuracies of 98.1% for drinking, 95.5% for grazing, and 49.3% for other behaviors, with an overall average recognition accuracy of 79.4%, representing a 7.4% improvement over the classic TSN model. Overall, the MFGN network effectively extracts and integrates global spatiotemporal features with fine motion features from cow video data, modeling both the overall sequence characteristics and focusing on local motion details, achieving precise behavioral recognition. This research not only enhances the accuracy of cow behavior recognition but also provides new technological means for precise management in modern smart agriculture, with broad industry application potential to improve farm efficiency, profitability, and disease control and prevention.

利用机器视觉技术识别奶牛行为在奶牛场的日常管理、健康监测、育种和繁殖中发挥着至关重要的作用,是现代智能农业的重要组成部分。本文介绍了一种用于分析奶牛视频数据的新型双流网络模型--运动聚焦全局局部网络(MFGN)。双流网络由全局时空特征流和精细运动特征流组成。全局时空特征流提取关键帧以去除冗余信息,并利用变换器网络进行全局时空特征提取,以反映奶牛视频的动态变化和视频帧之间的时空关系。精细运动特征流基于奶牛视频的帧差分,利用焦点卷积捕捉奶牛的细微运动,加强对细微行为变化的关注。为了评估所提出模型的性能,我们在澳大利亚的一个农场(CSIRO Armidale)收集了八头奶牛的视频数据样本,这些奶牛的身体和头部都做了标记,包括三个行为类别共 1715 个视频序列。该模型对喝水行为的识别准确率为 98.1%,对吃草行为的识别准确率为 95.5%,对其他行为的识别准确率为 49.3%,总体平均识别准确率为 79.4%,比传统的 TSN 模型提高了 7.4%。总之,MFGN 网络从奶牛视频数据中有效地提取并整合了全局时空特征和精细运动特征,既模拟了整体序列特征,又关注了局部运动细节,实现了精确的行为识别。这项研究不仅提高了奶牛行为识别的准确性,也为现代智慧农业的精准管理提供了新的技术手段,在提高农场效率、收益和疾病防控方面具有广阔的行业应用前景。
{"title":"Motion focus global–local network: Combining attention mechanism with micro action features for cow behavior recognition","authors":"","doi":"10.1016/j.compag.2024.109399","DOIUrl":"10.1016/j.compag.2024.109399","url":null,"abstract":"<div><p>The use of machine vision technology for recognizing cow behavior plays a crucial role in daily management, health monitoring, and breeding and reproduction in dairy farming, making it an essential component of modern smart agriculture. This paper presents a novel dual-stream network model, the Motion Focus Global-Local Network (MFGN), for analyzing cow video data. The dual-stream network consists of a global spatiotemporal feature stream and a fine motion feature stream. The global spatiotemporal feature stream extracts key frames to remove redundant information and utilizes a Transformer network for global spatio-temporal feature extraction, reflecting the dynamic changes in cow videos and the temporal relationships between video frames. The fine motion feature stream, based on frame differencing of cow videos, uses focal convolution to capture subtle movements of cows, enhancing the focus on minor behavioral changes. To evaluate the performance of the proposed model, video data samples were collected from eight cows marked on their bodies and heads at an Australian farm site (CSIRO Armidale), including a total of 1715 video sequences across three behavior categories. The model achieved recognition accuracies of 98.1% for drinking, 95.5% for grazing, and 49.3% for other behaviors, with an overall average recognition accuracy of 79.4%, representing a 7.4% improvement over the classic TSN model. Overall, the MFGN network effectively extracts and integrates global spatiotemporal features with fine motion features from cow video data, modeling both the overall sequence characteristics and focusing on local motion details, achieving precise behavioral recognition. This research not only enhances the accuracy of cow behavior recognition but also provides new technological means for precise management in modern smart agriculture, with broad industry application potential to improve farm efficiency, profitability, and disease control and prevention.</p></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":null,"pages":null},"PeriodicalIF":7.7,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142117445","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A detection method for dead caged hens based on improved YOLOv7 基于改进型 YOLOv7 的笼养死鸡检测方法
IF 7.7 1区 农林科学 Q1 AGRICULTURE, MULTIDISCIPLINARY Pub Date : 2024-08-31 DOI: 10.1016/j.compag.2024.109388

In large-scale laying hen farms, daily inspection of dead hens is a relevant task to monitor the health of the flock and prevent disease spreading. The current manual inspection method used in caged hen farms is inefficient, costly, and particularly difficult for high-rise cages. To address this issue, a dead caged hen detection method based on improved You Only Look Once version 7 (YOLOv7) was proposed in this study, which was optimized to improve detection performance and speed in complex farming environments, such as cage wire mesh occlusion and crowded hen occlusion. First, the Convolutional Block Attention Module was used to enable the model to learn target features accurately. Second, the Distance Intersection over Union Non-maximum Suppression and repulsion loss were introduced to improve crowded hen occlusion and reduce missed detections. Additionally, to facilitate the deployment of the proposed method on mobile devices, the MobileNetv3 lightweight network was used to replace the backbone of YOLOv7. Furthermore, the lightweight model was trained using the knowledge distillation method to enhance its performance. Finally, a comparison experiment of different object detection networks and an ablation experiment were conducted to evaluate the proposed method. The experimental results reveal that the improved YOLOv7 model proposed in this study performs optimally. Its precision, recall, F1 score, and [email protected] for the dead hens in the test set are 95.7 %, 86.8 %, 0.910, and 86.2 %, respectively. Compared with the original YOLOv7 model, precision, recall, and [email protected] were increased by 6 %, 10 %, and 13.4 %, respectively. The model parameters and Giga Floating-point Operations were decreased by 31.95 % and 60.56 %, respectively, resulting in a detection speed increase of 43 Frames Per Second. Furthermore, with the assistance of an inspection robot, the proposed dead hen detection model was deployed in the actual farming environments. Compared with methods proposed by other researchers, the proposed model is more suitable for complex actual farming environments and achieves higher detection accuracy, which can offer a reference for automated caged hen detection.

在大规模蛋鸡养殖场,每天检查死鸡是监测鸡群健康和防止疾病传播的一项重要任务。目前笼养鸡场使用的人工检查方法效率低、成本高,尤其是对于高层笼养鸡场而言更是困难重重。针对这一问题,本研究提出了一种基于改进型 You Only Look Once version 7(YOLOv7)的笼养死鸡检测方法,并对其进行了优化,以提高在笼养铁丝网遮挡和拥挤母鸡遮挡等复杂养殖环境下的检测性能和速度。首先,使用卷积块注意模块使模型能够准确地学习目标特征。其次,引入了 "联合非最大抑制距离交集 "和 "斥力损失",以改善拥挤的母鸡遮挡和减少漏检。此外,为了便于在移动设备上部署所提出的方法,使用了 MobileNetv3 轻量级网络来替代 YOLOv7 的骨干网络。此外,还使用知识提炼法对轻量级模型进行了训练,以提高其性能。最后,进行了不同物体检测网络的对比实验和烧蚀实验,以评估所提出的方法。实验结果表明,本研究提出的改进型 YOLOv7 模型性能最佳。对于测试集中的死母鸡,其精确度、召回率、F1 分数和 [email protected] 分别为 95.7 %、86.8 %、0.910 和 86.2 %。与最初的 YOLOv7 模型相比,精确度、召回率和 [email protected] 分别提高了 6%、10% 和 13.4%。模型参数和千兆浮点运算分别减少了 31.95 % 和 60.56 %,检测速度提高了 43 帧/秒。此外,在检测机器人的协助下,所提出的死鸡检测模型被部署在实际的养殖环境中。与其他研究人员提出的方法相比,所提出的模型更适合复杂的实际养殖环境,检测精度更高,可为笼养母鸡自动检测提供参考。
{"title":"A detection method for dead caged hens based on improved YOLOv7","authors":"","doi":"10.1016/j.compag.2024.109388","DOIUrl":"10.1016/j.compag.2024.109388","url":null,"abstract":"<div><p>In large-scale laying hen farms, daily inspection of dead hens is a relevant task to monitor the health of the flock and prevent disease spreading. The current manual inspection method used in caged hen farms is inefficient, costly, and particularly difficult for high-rise cages. To address this issue, a dead caged hen detection method based on improved You Only Look Once version 7 (YOLOv7) was proposed in this study, which was optimized to improve detection performance and speed in complex farming environments, such as cage wire mesh occlusion and crowded hen occlusion. First, the Convolutional Block Attention Module was used to enable the model to learn target features accurately. Second, the Distance Intersection over Union Non-maximum Suppression and repulsion loss were introduced to improve crowded hen occlusion and reduce missed detections. Additionally, to facilitate the deployment of the proposed method on mobile devices, the MobileNetv3 lightweight network was used to replace the backbone of YOLOv7. Furthermore, the lightweight model was trained using the knowledge distillation method to enhance its performance. Finally, a comparison experiment of different object detection networks and an ablation experiment were conducted to evaluate the proposed method. The experimental results reveal that the improved YOLOv7 model proposed in this study performs optimally. Its precision, recall, F1 score, and [email protected] for the dead hens in the test set are 95.7 %, 86.8 %, 0.910, and 86.2 %, respectively. Compared with the original YOLOv7 model, precision, recall, and [email protected] were increased by 6 %, 10 %, and 13.4 %, respectively. The model parameters and Giga Floating-point Operations were decreased by 31.95 % and 60.56 %, respectively, resulting in a detection speed increase of 43 Frames Per Second. Furthermore, with the assistance of an inspection robot, the proposed dead hen detection model was deployed in the actual farming environments. Compared with methods proposed by other researchers, the proposed model is more suitable for complex actual farming environments and achieves higher detection accuracy, which can offer a reference for automated caged hen detection.</p></div>","PeriodicalId":50627,"journal":{"name":"Computers and Electronics in Agriculture","volume":null,"pages":null},"PeriodicalIF":7.7,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142097259","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"农林科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computers and Electronics in Agriculture
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1