AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
Article Link
Collect
Submit Manuscript
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Research Article | Open Access

High-Throughput Spike Detection in Greenhouse Cultivated Grain Crops with Attention Mechanisms-Based Deep Learning Models

Sajid Ullah1,2,3Klára Panzarová3Martin Trtílek3Matej Lexa4Vojtěch Máčala4Kerstin Neumann5Thomas Altmann5Jan Hejátko1,2Markéta Pernisová1,2Evgeny Gladilin5( )
Mendel Centre for Plant Genomics and Proteomics, Central European Institute of Technology (CEITEC), Masaryk University, Brno, Czech Republic
National Centre for Biomolecular Research, Faculty of Science, Masaryk University, Brno, Czech Republic
Photon Systems Instruments, spol. s r.o., Drasov, Czech Republic
Faculty of Informatics, Masaryk University, Botanicka 68a, Brno, Czech Republic
Leibniz Institute of Plant Genetics and Crop Plant Research, Seeland OT Gatersleben, Germany
Show Author Information

Abstract

Detection of spikes is the first important step toward image-based quantitative assessment of crop yield. However, spikes of grain plants occupy only a tiny fraction of the image area and often emerge in the middle of the mass of plant leaves that exhibit similar colors to spike regions. Consequently, accurate detection of grain spikes renders, in general, a non-trivial task even for advanced, state-of-the-art deep neural networks (DNNs). To improve pattern detection in spikes, we propose architectural changes to Faster-RCNN (FRCNN) by reducing feature extraction layers and introducing a global attention module. The performance of our extended FRCNN-A vs. conventional FRCNN was compared on images of different European wheat cultivars, including “difficult” bushy phenotypes from 2 different phenotyping facilities and optical setups. Our experimental results show that introduced architectural adaptations in FRCNN-A helped to improve spike detection accuracy in inner regions. The mean average precision (mAP) of FRCNN and FRCNN-A on inner spikes is 76.0% and 81.0%, respectively, while on the state-of-the-art detection DNNs, Swin Transformer mAP is 83.0%. As a lightweight network, FRCNN-A is faster than FRCNN and Swin Transformer on both baseline and augmented training datasets. On the FastGAN augmented dataset, FRCNN achieved a mAP of 84.24%, FRCNN-A attained a mAP of 85.0%, and the Swin Transformer achieved a mAP of 89.45%. The increase in mAP of DNNs on the augmented datasets is proportional to the amount of the IPK original and augmented images. Overall, this study indicates a superior performance of attention mechanisms-based deep learning models in detecting small and subtle features of grain spikes.

References

1

Ullah S, Henke M, Narisetti N, Panzarová K, Trtílek M, Hejatko J, Gladilin E. Towards automated analysis of grain spikes in greenhouse images using neural network approaches: A comparative investigation of six methods. Sensors. 2021;21(22):7441.

2
Yang F, Choi W, Lin Y. Exploit all the layers: Fast and accurate CNN object detector with scale dependent pooling and cascaded rejection classifiers. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas (NV): IEEE; 2016. p. 2129–2137.
3

Bi K, Jiang P, Li L, Shi B, Wang C. Non-destructive measurement of wheat spike characteristics based on morphological image processing. Trans. Chin. Soc. Agric. Eng. 2010;26:212–216.

4

Misra T, Arora A, Marwaha S, Chinnusamy V, Rao AR, Jain R, Sahoo RN, Ray M, Kumar S, Raju D, et al. Spikesegnet—A deep learning approach utilizing encoder decoder network with hourglass for spike segmentation and counting in wheat plant from visual imaging. Plant Methods. 2020;16:40.

5
Pound MP, Atkinson JA, Wells DM, Pridmore TP, French AP. Deep learning for multitask plant phenotyping. In: Proceedings of the IEEE International Conference on Computer Vision Workshops (ICCVW). Venice (Italy): IEEE; 2017. p. 2055–2063.
6

Qiongyan L, Cai J, Berger B, Okamoto M, Miklavcic SJ. Detecting spikes of wheat plants using neural networks with Laws texture energy. Plant Methods. 2017;13:83.

7

Narisetti N, Neumann K, Röder MS, Gladilin E. Automated spike detection in diverse European wheat plants using textural features and the Frangi filter in 2D greenhouse images. Front Plant Sci. 2020;11:666.

8
Mottaghi R, Chen X, Liu X, Cho N-G, Lee S-W, Fidler S, Urtasun R, Tuille A, The role of context for object detection and semantic segmentation in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Columbus (OH): IEEE; 2014. p. 891–898.
9

Hasan MM, Chopin JP, Laga H, Miklavcic SJ. Detection and analysis of wheat spikes using convolutional neural networks. Plant Methods. 2018;14:100.

10
Girshick R, Donahue J, Darrell T, Malik J. Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Columbus (OH): IEEE; 2014. p. 580–587.
11

Ren S, He K, Girshick R, Sun J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015;39:1137–1149.

12

He K, Zhang X, Ren S, Sun J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans Pattern Anal Mach Intell. 2015;37:1904–1916.

13

Rahim UF, Mineno H. Tomato flower detection and counting in greenhouses using faster region-based convolutional neural network. J. Image Graph. 2020;8(4):107–113.

14
Hu J, Shen L, Sun G. Squeeze-and-excitation networks. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City (UT): IEEE; 2018. p. 7132–7141.
15
Guo J, Ma X, Sansom A, McGuire M, Kalaani A, Chen Q, Tang S, Yang Q, Fu S, Spanet: Spatial pyramid attention network for enhanced image recognition. In: Proceedings of the IEEE International Conference on Multimedia and Expo (ICME). London (UK): IEEE; 2020. p. 1–6.
16
Hu Y, Wen G, Luo M, Dai D, Ma J, and Yu Z. Competitive inner-imaging squeeze and excitation for residual network. arXiv. 2018. https://arxiv.org/abs/1807.08920
17
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Parmar N, Shazeer N, Vaswani A, Attention is all you need. Paper presented at: NIPS'17: Proceedings of the 31st International Conference on Neural Information Processing Systems; 2017 Dec 4; Long Beach, CA.
18
Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S. End-to-end object detection with transformers. In: European Conference on Computer Vision. Glasgow (UK): Springer; 2020. p. 213–229.
19
Tzutalin. LabelImg. 2015. https://github.com/tzutalin/labelImg
20
Liu B, Zhu Y, Song K, Elgammal A. Towards faster and stabilized gan training for high fidelity few-shot image synthesis. Paper presented at: International Conference on Learning Representations; 2021 May 3–7; Austria (virtual).
21
Seitzer M. pytorch-fid: FID Score for PyTorch. Version 0.3.0. 2020. https://github.com/mseitzer/pytorch-fid
22
He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas (NV): IEEE; 2016. p. 770–778.
23
Liu Z, Lin Y, Cao Y, Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. IEEE; 2021. p. 10012–10022.
Plant Phenomics
Article number: 0155
Cite this article:
Ullah S, Panzarová K, Trtílek M, et al. High-Throughput Spike Detection in Greenhouse Cultivated Grain Crops with Attention Mechanisms-Based Deep Learning Models. Plant Phenomics, 2024, 6: 0155. https://doi.org/10.34133/plantphenomics.0155

170

Views

0

Crossref

2

Web of Science

3

Scopus

0

CSCD

Altmetrics

Received: 21 July 2023
Accepted: 03 February 2024
Published: 11 March 2024
© 2024 Sajid Ullah et al. Exclusive licensee Nanjing Agricultural University. No claim to original U.S. Government Works.

Distributed under a Creative Commons Attribution License 4.0 (CC BY 4.0).

Return