Journal Home > Volume 3 , Issue 3

Based on the problems of high noise, lower contrast, and complex features in defaced images and the low accuracy of existing defaced image segmentation techniques, this paper proposes a defaced image segmentation algorithm based on DAC-CLGD-Danet. Firstly, a CBDNet asymmetric blind denoising network is used for noise-containing defaced images, and natural and synthetic images are trained together to model the image noise and enhance the denoising ability of natural noise. Secondly, Danet is used as the base network. A Dense Atrous Convolution module (DAC) is added to the dual attention mechanism module to extend the perceptual domain of deep convolution, reduce image feature loss, and enhance the representation of global information and edge features of defaced images; Cross-Level Gating Decoder module (CLGD) is introduced to lighten the segmentation network, enhance image context aggregation, and produce accurate semantic segmentation. The experimental results demonstrated that the method in this paper has a significant effect on the HRF dataset and Cityscapes dataset, with a significant improvement compared with FCN, UNet, and SETR models, with Intersection over Union (IoU) improved by 9.81% and Mean Intersection over Union (mIoU) improved by 3.01% compared with UNet.


menu
Abstract
Full text
Outline
About this article

A DAC-CLGD-Danet network based method for defaced image segmentation

Show Author's information Pengbo Li1Gang Li1Yibin He1Ling Zhang1( )Yuanjin Sun1Fayun Guo2
College of Software, Taiyuan University of Technology, Taiyuan 030024, China
Information Technology Department, Shanxi Taisen Technology Co. Ltd., Taiyuan 030082, China

Abstract

Based on the problems of high noise, lower contrast, and complex features in defaced images and the low accuracy of existing defaced image segmentation techniques, this paper proposes a defaced image segmentation algorithm based on DAC-CLGD-Danet. Firstly, a CBDNet asymmetric blind denoising network is used for noise-containing defaced images, and natural and synthetic images are trained together to model the image noise and enhance the denoising ability of natural noise. Secondly, Danet is used as the base network. A Dense Atrous Convolution module (DAC) is added to the dual attention mechanism module to extend the perceptual domain of deep convolution, reduce image feature loss, and enhance the representation of global information and edge features of defaced images; Cross-Level Gating Decoder module (CLGD) is introduced to lighten the segmentation network, enhance image context aggregation, and produce accurate semantic segmentation. The experimental results demonstrated that the method in this paper has a significant effect on the HRF dataset and Cityscapes dataset, with a significant improvement compared with FCN, UNet, and SETR models, with Intersection over Union (IoU) improved by 9.81% and Mean Intersection over Union (mIoU) improved by 3.01% compared with UNet.

Keywords: deep learning, neural networks, image segmentation, Danet, defaced images

References(36)

[1]

D. E. Alvarado-Carrillo and O. S. Dalmau-Cedeño, Width attention based convolutional neural network for retinal vessel segmentation, Expert Systems with Applications, vol. 209, p. 118313, 2022.

[2]

J. C. Liu, J. Xiang, L. Zhang, and G. Li, A self-constrained small sample defect image segmentation method, Journal of Chinese Computer Systems, vol. 43, no. 8, pp. 1732–1738, 2022.

[3]

A. Singh, G. S. Gaba, and M. Hedabou, Robust and effective image preprocessing conglomerate method for denoising of both grayscale and color images, Journal of Electronic Imaging, vol. 31, no. 4, p. 041203, 2021.

[4]

S. Chaudhary, S. Moon, and H. Lu, Fast, efficient, and accurate neuro-imaging denoising via supervised deep learning, Nature Communications, vol. 13, no. 1, p. 5165, 2022.

[5]

L. Zhang, J. Liu, F. Shang, G. Li, J. Zhao, and Y. Zhang, Robust segmentation method for noisy images based on an unsupervised denosing filter, Tsinghua Science and Technology, vol. 26, no. 5, pp. 736–748, 2021.

[6]
K. Yu, X. Wang, C. Dong, X. Tang, and C. C. Loy, Path-restore: Learning network path selection for image restoration, IEEE Transactions on Pattern Analysis and Machine Intelligence, doi: 10.1109/TPAMI.2021.3096255.
DOI
[7]

L. Deng, M. Yang, Z. Liang, Y. He, and C. Wang, Fusing geometrical and visual information via superpoints for the semantic segmentation of 3D road scenes, Tsinghua Science and Technology, vol. 25, no. 4, pp. 498–507, 2020.

[8]

L. Guo, L. Chen, C. L. P. Chen, and J. Zhou, Integrating guided filter into fuzzy clustering for noisy image segmentation, Digital Signal Processing, vol. 83, pp. 235–248, 2018.

[9]

Y. Wang, G. Yan, H. Zhu, S. Buch, Y. Wang, E. M. Haacke, J. Hua, and Z. Zhong, VC-net: Deep volume-composition networks for segmentation and visualization of highly sparse and noisy image data, IEEE Transactions on Visualization and Computer Graphics, vol. 27, no. 2, pp. 1301–1311, 2020.

[10]
V. Jain and H. S. Seung, Natural image denoising with convolutional networks, presented at Twenty-Second Annual Conference on Neural Information Processing Systems (NIPS), Vancouver, Canada, 2008.
[11]
H. C. Burger, C. J. Schuler, and S. Harmeling, Image denoising with multi-layer perceptrons, part 1: Comparison with existing algorithms and with bounds, arXiv preprint arXiv: 1211.1544, 2012.
[12]

K. Zhang, W. Zuo, Y. Chen, D. Meng, and L. Zhang, Beyond a Gaussian denoiser: Residual learning of deep CNN for image denoising, IEEE Transaction on Image Processing, vol. 26, no. 7, pp. 3142–3155, 2017.

[13]

D. Yang and J. Sun, BM3D-net: A convolutional neural network for transform-domain collaborative filtering, IEEE Signal Processing Letters, vol. 25, no. 1, pp. 55–59, 2017.

[14]
J. Jiang, L. Zheng, F. Luo, and Z. Zhang, RedNet: Residual encoder-decoder network for indoor RGB-D semantic segmentation, arXiv preprint arXiv: 1806.01054, 2018.
[15]
S. Guo, Z. Yan, K. Zhang, W. Zuo, and L. Zhang, Toward convolutional blind denoising of real photographs, in Proc. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 2019, pp. 1712–1722.
DOI
[16]

X. Zhang, B. Zhao, and G. Zhang, Improved parameter identification algorithm for ship model based on nonlinear innovation decorated by sigmoid function, Transportation Safety and Environment, vol. 3, no. 2, pp. 114–122, 2021.

[17]
J. Xu, Y. Huang, L. Liu, F. Zhu, and L. Shao, Noisy-as-clean: Learning unsupervised denoising from the corrupted image, arXiv preprint arXiv: 1906. 06878, 2020.
DOI
[18]
J. Long, E. Shelhamer, and T. Darrell, Fully convolutional networks for semantic tsegmentation, in Proc. IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 2015, pp. 3431–3440.
DOI
[19]
O. Ronneberger, P. Fischer, and T. Brox, U-net: Convolutional networks for biomedical image segmentation, in Proc. 18th International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 2015, pp. 234–241.
DOI
[20]

V. Badrinarayanan, A. Kendall, and R. Cipolla, SegNet: A deep convolutional encoder-decoder architecture for image segmentation, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 12, pp. 2481–2495, 2017.

[21]
L. C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuile, Semantic image segmentation with deep convolutional nets and fully connected CRFs, arXiv preprint arXiv: 1412.7062, 2016.
[22]

L. C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille, DeepLab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 40, no. 4, pp. 834–848, 2018.

[23]
L. C. Chen, G. Papandreou, F. Schroff, and H. Adam, Rethinking atrous convolution for semantic image segmentation, arXiv preprint arXiv: 1706.05587, 2017.
[24]
L. C. Chen, Y. Zhu, G. Papandreou, F. Schroff, and H. Adam, Encoder-decoder with atrous separable convolution for semantic image segmentation, in Proc. 15th European Conference on Computer Vision, Munich, Germany, 2018, 833–851.
DOI
[25]
J. Fu, J. Liu, H. Tian, Y. Li, Y. Bao, Z. Fang, and H. Lu, Dual attention network for scene segmentation, in Proc. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 2019, 3141–3149.
DOI
[26]
Y. Lu, Y. Chen, D. Zhao, and J. Chen, Graph-FCN for image semantic segmentation, in Proc. 16th International Symposium on Neural Networks, Moscow, Russia, 2019, pp. 97–105.
DOI
[27]
C. Yu, J. Wang, C. Peng, C. Gao, G. Yu, and N. Sang, BiseNet: Bilateral segmentation network for real-time semantic segmentation, in Proc. 15th European Conference on Computer Vision (ECCV), Munich, Germany, 2018, pp. 334–349.
DOI
[28]
Y. Wang, Q. Zhou, J. Liu, J. Xiong, G. Gao, X. Wu, and L. J. Latecki, Lednet: A lightweight encoder-decoder network for real-time semantic segmentation, in Proc. 2019 IEEE International Conference on Image Processing (ICIP), Taiwan, Province of China, 2019, pp. 1860–1864.
DOI
[29]

R. Niu, X. Sun, Y. Tian, W. Diao, K. Chen, and K. Fu, Hybrid multiple attention network for semantic segmentation in aerial images, IEEE Transactions on Geoscience and Remote Sensing, vol. 60, p. 5603018, 2021.

[30]

E. Xie, W. Wang, Z. Yu, A. Anandkumar, J. M. Alvarez, and P. Luo, SegFormer: Simple and efficient design for semantic segmentation with transformers, Advances in Neural Information Processing Systems, vol. 34, pp. 12077–12090, 2021.

[31]
T. Meinhardt, A. Kirillov, L. Leal-Taixe, and C. Feichtenhofer, Trackformer: Multi-object tracking with transformers, in Proc. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 2022, pp. 8834–8844.
DOI
[32]
X. Chen, B. Yan, J. Zhu, D. Wang, X. Yang, and H. Lu, Transformer tracking, in Proc. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 2021, pp. 8122–8133.
DOI
[33]
S. He, H. Luo, P. Wang, F. Wang, H. Li, and W. Jiang, TransReID: Transformer-based object re-identification, in Proc. 2021 IEEE/CVF International Conference on Computer Vision, Montreal, Canada, 2021, pp. 14993–15002.
DOI
[34]

Z. Gu, J. Cheng, H. Fu, K. Zhou, H. Hao, Y. Zhao, T. Zhang, S. Gao, and J. Liu, Ce-net: Context encoder network for 2D medical image segmentation, IEEE Transactions on Medical Imaging, vol. 38, no. 10, pp. 2281–2292, 2019.

[35]

S. A. Taghanaki, K. Abhishek, J. P. Cohen, J. Cohen-Adad, and G. Hamarneh, Deep semantic segmentation of natural and medical images: A review, Artificial Intelligence Review, vol. 54, no. 1, pp. 137–178, 2021.

[36]

J. Fu, J. Liu, J. Jiang, Y. Li, Y. Bao, and H. Lu, Scene segmentation with dual relation-aware attention network, IEEE Transactions on Neural Networks and Learning Systems, vol. 32, no. 6, pp. 2547–2560, 2020.

Publication history
Copyright
Acknowledgements
Rights and permissions

Publication history

Received: 27 October 2022
Revised: 24 November 2022
Accepted: 02 December 2022
Published: 30 September 2022
Issue date: September 2022

Copyright

© All articles included in the journal are copyrighted to the ITU and TUP.

Acknowledgements

Acknowledgment

This work was supported by the Central Leading Local Special Foundation of Shanxi Province (Nos. YDZJSX2021C004 and YDZJSX2022A016), and the Natural Science Foundation of Shanxi Province (No. 20210302124554);

Rights and permissions

This work is available under the CC BY-NC-ND 3.0 IGO license:https://creativecommons.org/licenses/by-nc-nd/3.0/igo/

Return