AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
Article Link
Collect
Submit Manuscript
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Research Article | Open Access

Awn Image Analysis and Phenotyping Using BarbNet

Narendra Narisetti1( )Muhammad Awais2Muhammad Khan2Frieder Stolzenburg3Nils Stein2,4Evgeny Gladilin1
Leibniz Institute for Plant Genetics and Crop Plant Research (IPK), Molecular Genetics, 06466 Seeland, Germany
Leibniz Institute for Plant Genetics and Crop Plant Research (IPK), Genebank, 06466 Seeland, Germany
Harz University of Applied Sciences, Automation and Computer Sciences Department, 38855 Wernigerode, Germany
Center of integrated Breeding Research (CiBreed), Department of Crop Sciences, Georg-August-University, 37075 Göttingen, Germany
Show Author Information

Abstract

Consideration of the properties of awns is important for the phenotypic description of grain crops. Awns have a number of important functions in grasses, including assimilation, mechanical protection, and seed dispersal and burial. An important feature of the awn is the presence or absence of barbs—tiny hook-like single-celled trichomes on the outer awn surface that can be visualized using microscopic imaging. There are, however, no suitable software tools for the automated analysis of these small, semi-transparent structures in a high-throughput manner. Furthermore, automated analysis of barbs using conventional methods of pattern detection and segmentation is hampered by high variability of their optical appearance including size, shape, and surface density. In this work, we present a software tool for automated detection and phenotyping of barbs in microscopic images of awns, which is based on a dedicated deep learning model (BarbNet). Our experimental results show that BarbNet is capable of detecting barb structures in different awn phenotypes with an average accuracy of 90%. Furthermore, we demonstrate that phenotypic traits derived from BarbNet-segmented images enable a quite robust categorization of 4 contrasting awn phenotypes with an accuracy of >85%. Based on the promising results of this work, we see that the proposed model has potential applications in the automation of barley awns sorting for plant developmental analysis.

References

1

Abebe T, Wise RP, Skadsen RW. Comparative transcriptional profiling established the awn as the major photosynthetic organ of the barley spike while the lemma and the Palea primarily protect the seed. Plant Genome. 2009;2(3):0019.

2

Elbaum R, Zaltzman L, Burgert I, Fratzl P. The role of wheat awns in the seed dispersal unit. Science. 2007;316(5823):884–886.

3

Kondorosi E, Roudier F, Gendreau E. Plant cell-size control: Growing by ploidy? Curr Opin Plant Biol. 2000;3(6):488–492.

4

Milner SG, Jost M, Taketa S, Mazón ER, Himmelbach A, Oppermann M, Weise S, Knüpffer H, Basterrechea M, König P, et al. Genebank genomics highlights the diversity of a global barley collection. Nat Genet. 2019;51(2):319–326.

5

Huang C, Becker MF, Keto JW, Kovar D. Annealing of nanostructured silver films produced by supersonic deposition of nanoparticles. J Appl Phys. 2007;102(5):054308.

6
The GIMP Development Team. Gimp, version 2.10.12, 12 Jun 2019. https://www.gimp.org
7

Al-Kofahi Y, Zaltsman A, Graves R, Marshall W, Rusu M. A deep learning-based algorithm for 2-d cell segmentation in microscopy images. BMC bioinformatics. 2018;1:365.

8
Narotamo H, Sanches JM, Silveira M. Segmentation of cell nuclei in fluorescence microscopy images using deep learning. Paper presented at: Pattern Recognition and Image Analysis: 9th Iberian Conference, IbPRIA 2019; 2019 Jul 1–4; Madrid, Spain.
9

Sekh AA, Opstad IS, Godtliebsen G, Birgisdottir ÅB, Ahluwalia BS, Agarwal K, Prasad DK. Physics-based machine learning for subcellular segmentation in living cells. Nat Mach Intell. 2021;3(10):1071.

10

Wang T, Rostamza M, Song Z, Wang L, McNickle G, Iyer-Pascuzzi AS, Qiu Z, Jin J. Segroot: A high throughput segmentation method for root image analysis. Comput Electron Agric. 2019;162:845–854.

11

Narisetti N, Henke M, Seiler C, Shi R, Junker A, Altmann T, Gladilin E. Semi-automated root image analysis (saRIA). Sci Rep. 2019;9(1):19674.

12

Narisetti N, Henke M, Seiler C, Junker A, Ostermann J, Altmann T, Gladilin E. Fully-automated root image analysis (faRIA). Sci Rep. 2021;11(1):16047.

13

Rakhmatulin I, Kamilaris A, Andreasen C. Deep neural networks to detect weeds from crops in agricultural environments in real-time: A review. Remote Sens. 2021;13(21):4486.

14

Narisetti N, Henke M, Neumann K, Stolzenburg F, Altmann T, Gladilin E. Deep learning based greenhouse image segmentation and shoot phenotyping (deepshoot). Front Plant Sci. 2022;13:906410.

15
Hu K. Deep learning techniques for in-crop weed identification: A review. ArXiv 2021. preprint arXiv: 2103.14872.
16

Ullah S, Henke M, Narisetti N, Panzarová K, Trtílek M, Hejatko J, Gladilin E. Towards automated analysis of grain spikes in greenhouse images using neural network approaches: A comparative investigation of six methods. Sensors. 2021;21(22):7441.

17

Misra T, Arora A, Marwaha S, Chinnusamy V, Rao AR, Jain R, Sahoo RN, Ray M, Kumar S, Raju D, et al. Spikesegnet—A deep learning approach utilizing encoder-decoder network with hourglass for spike segmentation and counting in wheat plant from visual imaging. Plant Methods. 2020;16:40.

18
Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. In: International Conference on Medical image computing and computer-assisted intervention. Munich (Germany): Springer; 2015. p. 234–241.
19
Ioffe S, Szegedy C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. ArXiv. 2015. https://doi.org/10.48550/arXiv.1502.03167
20
Santurkar S, Tsipras D, Ilyas A, Madry A. How does batch normalization help optimization. ArXiv. 2019. https://doi.org/10.48550/arXiv.1805.11604
21
Li X, Chen S, Hu X. J. Yang. Understanding the disharmony between dropout and batch normalization by variance shift. ArXiv. 2018. https://doi.org/10.48550/arXiv.1801.05134
22
Peng C, Zhang X, Yu G, Luo G, Sun J. Large kernel matters—Improve semantic segmentation by global convolutional network. Paper presented at: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2017 Jul 21–26; Hawaii.
23
Agostinelli F, Hoffman M, Sadowski P, Baldi P. Learning activation functions to improve deep neural networks. ArXiv. 2014. https://doi.org/10.48550/arXiv.1412.6830.
24
Wang L, Guo S, Huang W, Qiao Y. Places205-vggnet models for scene recognition. ArXiv. 2015. https://doi.org/10.48550/arXiv.1508.01667
25

Jha RR, Jaswal G, Gupta D, Saini S, Nigam A. Pixisegnet: Pixel-level iris segmentation network using convolutional encoder-decoder with stacked hourglass bottleneck. IET Biometrics. 2019;9(1):11–24.

26
Dunne RA, Campbell NA. On the pairing of the softmax activation and cross-entropy penalty functions and the derivation of the softmax activation function. Paper presented at: Proc. 8th Aust. Conf. on the Neural Networks, Melbourne, Australia: Citeseer; 1997.
27

Zou KH, Warfield SK, Bharatha A, Tempany CMC, Kaus MR, Haker SJ, Wells WM III, Jolesz FA, Kikinis R. Statistical validation of image segmentation quality based on a spatial overlap index1: Scientific reports. Acad Radiol. 2004;11(2):178–189.

28
Abadi M, Agarwal A, Barham P, Brevdo E, Chen Z, Citro C, Corrado GS, Davis A, Dean J, Devin M, et al. Tensorflow: Large-scale machine learning on heterogeneous distributed systems. ArXiv. 2016. https://doi.org/10.48550/arXiv.1603.04467
29

van de Warlt S, Colbert SC, Varoquaux G. The numpy array: A structure for efficient numerical computation. Comput Sci Eng. 2011;13(2):22–30.

30

Van der Walt S, Schonberger JL, Nunez-Iglesias J, Boulogne F, Warner JD, Yager N, Gouillart E, Yu T. Scikit-image: Image processing in python. PeerJ. 2014;2:e453.

31
Crimi A, Bakas S, Kuijf H, Menze B, Reyes M. Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: Third International Workshop, BrainLes 2017, Held in Conjunction with MICCAI 2017, Quebec City, QC, Canada, September 14, 2017, Revised Selected Papers; Springer; 2018.
32
Joseph VR. Optimal ratio for data splitting. Statistical Analysis and Data Mining: The ASA Data Science Journal. 2022.
33
Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. In: Advances in neural information processing systems. Cambridge (MA): Massachusetts Institute of Technology Press; 2012. p. 1097–1105.
34
Kingma DP, Ba J. Adam: A method for stochastic optimization. ArXiv 2014. https://doi.org/10.48550/arXiv.1412.6980
35
Bengio Y. Practical recommendations for gradient-based training of deep architectures. In: Neural networks: Tricks of the trade. Heidelberg (Germany): Springer; 2012. p. 437–478.
Plant Phenomics
Article number: 0081
Cite this article:
Narisetti N, Awais M, Khan M, et al. Awn Image Analysis and Phenotyping Using BarbNet. Plant Phenomics, 2023, 5: 0081. https://doi.org/10.34133/plantphenomics.0081

137

Views

1

Crossref

1

Web of Science

1

Scopus

0

CSCD

Altmetrics

Received: 13 April 2023
Accepted: 23 July 2023
Published: 04 August 2023
© 2023 Narendra Narisetti et al. Exclusive licensee Nanjing Agricultural University. No claim to original U.S. Government Works.

Distributed under a Creative Commons Attribution License (CC BY 4.0).

Return