Connor C. Mullins, Travis J. Esau, Qamar U. Zaman, Patrick J. Hennessy
{"title":"Optimizing data collection requirements for machine learning models in wild blueberry automation through the application of DALL-E 2","authors":"Connor C. Mullins, Travis J. Esau, Qamar U. Zaman, Patrick J. Hennessy","doi":"10.1016/j.atech.2024.100764","DOIUrl":null,"url":null,"abstract":"<div><div>This research developed a workflow to assess the viability of AI-generated imagery in training machine learning models for detecting ripe wild blueberries (<em>Vaccinium angustifolium</em> Ait.), hair fescue weeds (<em>Festuca filiformis</em> Pourr.), and red leaf disease (<em>Exobasidium vaccinii</em>). Ground truth images were collected and augmented with AI-generated variations using DALL-E 2 to expand the dataset. Models were trained on three datasets: ground truth, generated, and a combination (40% generated images). Evaluation metrics included precision, recall, mAP<sub>50</sub>, and mAP<sub>50–95</sub>, analyzed using ANOVA multiple mean comparisons and Tukey's HSD test (α = 0.05). For ripe wild blueberries, combination models achieved the highest performance across all metrics (mAP<sub>50</sub>: 0.834), significantly outperforming the ground truth model (mAP<sub>50</sub>: 0.806) in terms of mAP<sub>50–95</sub> (0.478 compared to 0.424). For hair fescue weeds, the combination dataset outperformed others with the highest mAP<sub>50</sub> (0.983), closely followed by the ground truth dataset (mAP<sub>50</sub>: 0.969). In detecting red leaf disease, the combination dataset showed the best performance (mAP<sub>50</sub>: 0.848 ± 0.140, mAP<sub>50–95</sub>: 0.607 ± 0.219), compared to the ground truth (mAP<sub>50</sub>: 0.615 ± 0.092, mAP<sub>50–95</sub>: 0.417 ± 0.045) and generated datasets (mAP<sub>50</sub>: 0.245 ± 0.088, mAP<sub>50–95</sub>: 0.144 ± 0.059). Models trained solely on generated images showed significantly lower performance across all categories except the precision metric for red leaf, where performance was comparable to ground truth. This indicated that while AI-generated images can augment datasets and improve generalization, they cannot fully replace ground truth data while maintaining model performance. Integrating AI-generated images with real-world data significantly improved model performance, reduced labor-intensive data collection processes, and provided a more diverse and comprehensive dataset for training, underscoring the importance of a balanced approach to optimizing data collection protocols for wild blueberry cultivation.</div></div>","PeriodicalId":74813,"journal":{"name":"Smart agricultural technology","volume":"10 ","pages":"Article 100764"},"PeriodicalIF":6.3000,"publicationDate":"2024-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Smart agricultural technology","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S277237552400368X","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AGRICULTURAL ENGINEERING","Score":null,"Total":0}
引用次数: 0
Abstract
This research developed a workflow to assess the viability of AI-generated imagery in training machine learning models for detecting ripe wild blueberries (Vaccinium angustifolium Ait.), hair fescue weeds (Festuca filiformis Pourr.), and red leaf disease (Exobasidium vaccinii). Ground truth images were collected and augmented with AI-generated variations using DALL-E 2 to expand the dataset. Models were trained on three datasets: ground truth, generated, and a combination (40% generated images). Evaluation metrics included precision, recall, mAP50, and mAP50–95, analyzed using ANOVA multiple mean comparisons and Tukey's HSD test (α = 0.05). For ripe wild blueberries, combination models achieved the highest performance across all metrics (mAP50: 0.834), significantly outperforming the ground truth model (mAP50: 0.806) in terms of mAP50–95 (0.478 compared to 0.424). For hair fescue weeds, the combination dataset outperformed others with the highest mAP50 (0.983), closely followed by the ground truth dataset (mAP50: 0.969). In detecting red leaf disease, the combination dataset showed the best performance (mAP50: 0.848 ± 0.140, mAP50–95: 0.607 ± 0.219), compared to the ground truth (mAP50: 0.615 ± 0.092, mAP50–95: 0.417 ± 0.045) and generated datasets (mAP50: 0.245 ± 0.088, mAP50–95: 0.144 ± 0.059). Models trained solely on generated images showed significantly lower performance across all categories except the precision metric for red leaf, where performance was comparable to ground truth. This indicated that while AI-generated images can augment datasets and improve generalization, they cannot fully replace ground truth data while maintaining model performance. Integrating AI-generated images with real-world data significantly improved model performance, reduced labor-intensive data collection processes, and provided a more diverse and comprehensive dataset for training, underscoring the importance of a balanced approach to optimizing data collection protocols for wild blueberry cultivation.