Skip to main content

2024 | OriginalPaper | Buchkapitel

Enhancing YOLOv7 for Plant Organs Detection Using Attention-Gate Mechanism

verfasst von : Hanane Ariouat, Youcef Sklab, Marc Pignal, Florian Jabbour, Régine Vignes Lebbe, Edi Prifti, Jean-Daniel Zucker, Eric Chenin

Erschienen in: Advances in Knowledge Discovery and Data Mining

Verlag: Springer Nature Singapore

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Herbarium scans are valuable raw data for studying how plants adapt to climate change and respond to various factors. Characterization of plant traits from these images is important for investigating such questions, thereby supporting plant taxonomy and biodiversity description. However, processing these images for meaningful data extraction is challenging due to scale variance, complex backgrounds that contain annotations, and the variability in specimen color, shape, and orientation of specimens. In addition, the plant organs often appear compressed, deformed, or damaged, with overlapping occurrences that are common in scans. Traditional organ recognition techniques, while adept at segmenting discernible plant characteristics, are limited in herbarium scanning applications. Two automated methods for plant organ identification have been previously reported. However, they show limited effectiveness, especially for small organs. In this study we introduce YOLOv7-ag model, which is a novel model based on the YOLOv7 that incorporates an attention-gate mechanism, which enhances the detection of plant organs, including stems, leaves, flowers, fruits, and seeds. YOLOv7-ag significantly outperforms previous state of the art as well as the original YOLOv7 and YOLOv8 models with a precision and recall rate of 99.2% and 98.0%, respectively, particularly in identifying small plant organs.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Fußnoten
1
After the backbone network and before the detection layer.
 
2
The three scales are designed to detect small, medium, and large objects, respectively. They are represented by feature maps that are extracted at different depths of the neural network, thus allowing for precise detection across a varied range of object sizes.
 
3
Ground truth represents the desired bounding box as output of the object detection algorithm.
 
Literatur
1.
Zurück zum Zitat Ariouat, H., et al.: Extracting masks from herbarium specimen images based on object detection and image segmentation techniques. Biodiv. Inf. Sci. Stand. 7, e112161 (2023) Ariouat, H., et al.: Extracting masks from herbarium specimen images based on object detection and image segmentation techniques. Biodiv. Inf. Sci. Stand. 7, e112161 (2023)
2.
Zurück zum Zitat Sahraoui, M., Sklab, Y., Pignal, M., Lebbe, R.V., Guigue, V.: Leveraging Multimodality for Biodiversity Data: exploring joint representations of species descriptions and specimen images using CLIP. Biodivers. Inf. Sci. Stand. 7(2023), e112666 (2023) Sahraoui, M., Sklab, Y., Pignal, M., Lebbe, R.V., Guigue, V.: Leveraging Multimodality for Biodiversity Data: exploring joint representations of species descriptions and specimen images using CLIP. Biodivers. Inf. Sci. Stand. 7(2023), e112666 (2023)
3.
Zurück zum Zitat Meredith, L.: Roles of natural history collections. Ann. Mo. Bot. Gard. 4(83), 536–545 (1996) Meredith, L.: Roles of natural history collections. Ann. Mo. Bot. Gard. 4(83), 536–545 (1996)
4.
Zurück zum Zitat Raven, P.H.: Saving plants, saving ourselves. Plants People Planet 1, 8–13 (2019)CrossRef Raven, P.H.: Saving plants, saving ourselves. Plants People Planet 1, 8–13 (2019)CrossRef
5.
Zurück zum Zitat Younis, S., Schmidt, M., Weiland, C., Dressler, S., Seeger, B., Hickler, T.: Detection and annotation of plant organs from digitised herbarium scans using deep learning. Biodiv. Data J. 8, e57090 (2020)CrossRef Younis, S., Schmidt, M., Weiland, C., Dressler, S., Seeger, B., Hickler, T.: Detection and annotation of plant organs from digitised herbarium scans using deep learning. Biodiv. Data J. 8, e57090 (2020)CrossRef
6.
Zurück zum Zitat Besnard, G., et al.: Herbarium-based science in the twenty-first century. Botany Lett. 165, 323–327 (2018)CrossRef Besnard, G., et al.: Herbarium-based science in the twenty-first century. Botany Lett. 165, 323–327 (2018)CrossRef
7.
Zurück zum Zitat Soltis, P.: Digitization of herbaria enables novel research. Am. J. Bot. 104, 1281–1284 (2017)CrossRef Soltis, P.: Digitization of herbaria enables novel research. Am. J. Bot. 104, 1281–1284 (2017)CrossRef
8.
Zurück zum Zitat Abdelaziz, T., Bassem, B., Walid, M.: A deep learning-based approach for detecting plant organs from digitized herbarium specimen images. Eco. Inform. 69, 101590 (2022)CrossRef Abdelaziz, T., Bassem, B., Walid, M.: A deep learning-based approach for detecting plant organs from digitized herbarium specimen images. Eco. Inform. 69, 101590 (2022)CrossRef
9.
Zurück zum Zitat Patrick, W., et al.: Large-scale digitization of herbarium specimens: development and usage of an automated, high-throughput conveyor system. Taxon 67, 165–178 (2018)CrossRef Patrick, W., et al.: Large-scale digitization of herbarium specimens: development and usage of an automated, high-throughput conveyor system. Taxon 67, 165–178 (2018)CrossRef
10.
Zurück zum Zitat Wenli, Z., et al.: Deep-learning-based in-field citrus fruit detection and tracking. Horticult. Res. 9, uhac003 (2022) Wenli, Z., et al.: Deep-learning-based in-field citrus fruit detection and tracking. Horticult. Res. 9, uhac003 (2022)
11.
Zurück zum Zitat Jiang, Y., Li, C.: Convolutional neural networks for image-based high-throughput plant phenotyping: a review. Plant Phenom. 9 (2020) Jiang, Y., Li, C.: Convolutional neural networks for image-based high-throughput plant phenotyping: a review. Plant Phenom. 9 (2020)
12.
Zurück zum Zitat Borhani, Y., Khoramdel, J., Najafi, E.: A deep learning based approach for automated plant disease classification using vision transformer. Sci. Rep. 12, 11554 (2022)CrossRef Borhani, Y., Khoramdel, J., Najafi, E.: A deep learning based approach for automated plant disease classification using vision transformer. Sci. Rep. 12, 11554 (2022)CrossRef
13.
Zurück zum Zitat Ashish, V., et al.: Attention is all you need. CoRR abs/1706.03762 (2017) Ashish, V., et al.: Attention is all you need. CoRR abs/1706.03762 (2017)
14.
Zurück zum Zitat Sue Han, L., Chee Seng, C., Simon, J., Paolo, R.: How deep learning extracts and learns leaf features for plant classification. Pattern Recogn. 71, 1–13 (2017)CrossRef Sue Han, L., Chee Seng, C., Simon, J., Paolo, R.: How deep learning extracts and learns leaf features for plant classification. Pattern Recogn. 71, 1–13 (2017)CrossRef
15.
Zurück zum Zitat Mochida, K., et al.: Computer vision-based phenotyping for improvement of plant productivity: a machine learning perspective. GigaScience 8, giy153 (2018) Mochida, K., et al.: Computer vision-based phenotyping for improvement of plant productivity: a machine learning perspective. GigaScience 8, giy153 (2018)
16.
Zurück zum Zitat Shaoqing, R., Kaiming, H., Ross, B.G., Jian, S.: Faster R-CNN: towards real-time object detection with region proposal networks. Computer Vision and Pattern Recognition abs/1506.01497 (2015) Shaoqing, R., Kaiming, H., Ross, B.G., Jian, S.: Faster R-CNN: towards real-time object detection with region proposal networks. Computer Vision and Pattern Recognition abs/1506.01497 (2015)
17.
Zurück zum Zitat Kaiming, H., Georgia, G., Piotr, D., Ross, B.G.: Mask R-CNN. Computer Vision and Pattern Recognition, vol. abs/1703.06870 (2017) Kaiming, H., Georgia, G., Piotr, D., Ross, B.G.: Mask R-CNN. Computer Vision and Pattern Recognition, vol. abs/1703.06870 (2017)
18.
Zurück zum Zitat Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016) Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016)
19.
Zurück zum Zitat Xiao, Z., Lang, J., Shuai, L., Tingting, Z., Xingang, M.: YOLO-SASE: an improved YOLO algorithm for the small targets detection in complex backgrounds. Computer Vision and Pattern Recognition, vol. abs/2207.02696 (2022) Xiao, Z., Lang, J., Shuai, L., Tingting, Z., Xingang, M.: YOLO-SASE: an improved YOLO algorithm for the small targets detection in complex backgrounds. Computer Vision and Pattern Recognition, vol. abs/2207.02696 (2022)
20.
Zurück zum Zitat Chien-Yao, W., Alexey, B., Hong-Yuan Mark, L.: YOLOv7: trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. Computer Vision and Pattern Recognition, vol. abs/2207.02696 (2022) Chien-Yao, W., Alexey, B., Hong-Yuan Mark, L.: YOLOv7: trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. Computer Vision and Pattern Recognition, vol. abs/2207.02696 (2022)
21.
Zurück zum Zitat Ozan, O., et al.: Attention U-Net: learning where to look for the pancreas. Computer Vision and Pattern Recognition, vol. abs/1804.03999 (2018) Ozan, O., et al.: Attention U-Net: learning where to look for the pancreas. Computer Vision and Pattern Recognition, vol. abs/1804.03999 (2018)
22.
Zurück zum Zitat Lang, P.M., Willems, F., Scheepens, J.F., Burbano, H., Bossdorf, O.: Using herbaria to study global environmental change. New Phytol. 2021, 110–122 (2019)CrossRef Lang, P.M., Willems, F., Scheepens, J.F., Burbano, H., Bossdorf, O.: Using herbaria to study global environmental change. New Phytol. 2021, 110–122 (2019)CrossRef
23.
Zurück zum Zitat Zhao, H., Zhang, H., Zhao, Y.: YOLOv7-Sea: object detection of maritime UAV images based on improved YOLOv7. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) Workshops (2023) Zhao, H., Zhang, H., Zhao, Y.: YOLOv7-Sea: object detection of maritime UAV images based on improved YOLOv7. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) Workshops (2023)
24.
Zurück zum Zitat Zixiao, Z., et al.: ViT-YOLO: transformer-based YOLO for object detection. In: IEEE/CVF International Conference on Computer Vision Workshops, ICCVW 2021, Montreal, BC, Canada, 11–17 October 2021 Zixiao, Z., et al.: ViT-YOLO: transformer-based YOLO for object detection. In: IEEE/CVF International Conference on Computer Vision Workshops, ICCVW 2021, Montreal, BC, Canada, 11–17 October 2021
25.
Zurück zum Zitat James, C., et al.: From prototype to inference: a pipeline to apply deep learning in sorghum panicle detection. Plant Phenomics 5, 0017 (2023)CrossRef James, C., et al.: From prototype to inference: a pipeline to apply deep learning in sorghum panicle detection. Plant Phenomics 5, 0017 (2023)CrossRef
26.
Zurück zum Zitat Jie, X., et al.: TrichomeYOLO: a neural network for automatic maize trichome counting. Plant Phenom. 5, 0024 (2023)CrossRef Jie, X., et al.: TrichomeYOLO: a neural network for automatic maize trichome counting. Plant Phenom. 5, 0024 (2023)CrossRef
27.
Zurück zum Zitat Zhaoyang, N., Guoqiang, Z., Hui, Y.: A review on the attention mechanism of deep learning. Plant Neurocomput. 452, 48–62 (2021)CrossRef Zhaoyang, N., Guoqiang, Z., Hui, Y.: A review on the attention mechanism of deep learning. Plant Neurocomput. 452, 48–62 (2021)CrossRef
28.
Zurück zum Zitat Dillon, R., Jordan, K., Jacqueline, H., Ahmad, D.: Real-time flying object detection with YOLOv8. Computer Vision and Pattern Recognition (2023) Dillon, R., Jordan, K., Jacqueline, H., Ahmad, D.: Real-time flying object detection with YOLOv8. Computer Vision and Pattern Recognition (2023)
Metadaten
Titel
Enhancing YOLOv7 for Plant Organs Detection Using Attention-Gate Mechanism
verfasst von
Hanane Ariouat
Youcef Sklab
Marc Pignal
Florian Jabbour
Régine Vignes Lebbe
Edi Prifti
Jean-Daniel Zucker
Eric Chenin
Copyright-Jahr
2024
Verlag
Springer Nature Singapore
DOI
https://doi.org/10.1007/978-981-97-2253-2_18

Premium Partner