AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
PDF (15.5 MB)
Collect
Submit Manuscript AI Chat Paper
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Publishing Language: Chinese

Online intelligent positioning of parts based on adaptive inspired particle swarm optimization

Huasen LIU1Zihao GUO2( )Haiping NIE1Zhijun PEN1
AVIC Chengdu Aircraft Industrial (Group) Co., Ltd., Chengdu 610092, China
Tianjin Research Institute for Advanced Equipment, Tsinghua University, Tianjin 300300, China
Show Author Information

Abstract

Objective

In the realm of industrial assembly line production, the automation of spraying operations for small- and medium-sized parts often encounters significant challenges. The primary issue lies in the inconsistent placement of different parts, complicating the automation of measurement and positioning tasks.

Methods

To address this, a sophisticated approach leveraging 2D images has been developed to assist in the preliminary positioning and classification of parts. This method involves planning the scanning and measurement path of a 3D camera based on the preliminary positioning information provided by 2D image processing. However, the cloud data points captured by the 3D camera include a large amount of irrelevant background information. To isolate accurate point cloud data for the parts, the data is segmented according to the preliminary positioning information of the parts. Point cloud registration, a critical challenge within robotics and computer vision fields, involves estimating a rigid transformation to align one point cloud to another. The final step entails registering the measured point cloud with the 3D digital model of the part to determine the exact position and orientation of the part. The 2D image is processed successively by filtering, gray equalization, binarization, morphological processing, and contour extraction. These steps effectively separate the image foreground from the background and identify the parts within the image. Recently, an innovative adaptation of the classic particle swarm optimization algorithm, enhanced with an adaptive heuristic, has been employed to tailor different schemes and scales based on specific registration conditions. This approach achieves rapid location cloud registration under strong background noise. To overcome issues of slow convergence speed and the tendency to settle into local optima, this improved algorithm incorporates learning and inertia coefficients of adaptive state registration, along with stalling coefficients and gradient descent operations for adaptive scaling. Employing a normal distribution confidence criterion minimizes the effect of fitness outliers on registration, facilitating intelligent alignment between the point cloud and the theoretical numerical model. This helps achieve precise determination of the parts' positions and orientations.

Results

Integrating two-dimensional (2D) vision technology significantly reduces measurement times and improves efficiency through multithreaded concurrent synchronization operations. Finally, the automatic scanning of three-dimensional (3D) point clouds and the autonomous registration positioning of various parts are accomplished within 3 min, achieving an average accuracy of 2 mm. When aligning identical point clouds, the algorithm demonstrates a registration accuracy of 0.002 mm.

Conclusions

Despite its robustness under strong back-point cloud influence, the algorithm's registration accuracy is still greatly affected. In addition, the inherent discrepancy between the sampling consistency of the scanned point cloud and the theoretical numerical model introduces an error of about 0.5 mm, limiting further improvements in registration accuracy. Due to their limited features and small size, some aviation parts are easily misidentified and prone to significant attitude registration errors. It is necessary to further combine the advantages of 2D images and 3D point cloud data to enhance the robustness of the identification and positioning of aviation parts.

CLC number: TP391 Document code: A Article ID: 1000-0054(2024)08-1367-13

References

[1]

WANG R Y, TANG L W, TANG T. Fast sample adaptive offset jointly based on HOG features and depth information for VVC in visual sensor networks [J]. Sensors, 2020, 20(23): 6754.

[2]

ZHANG H, WANG K F, WANG F Y. Advances and perspectives on applications of deep learning in visual object detection [J]. Acta Automatica Sinica, 2017, 43(8): 1289-1305. (in Chinese)

[3]

ZHAO J, ZHAO J, ZHANG L, et al. Vision algorithm and experimental study on the robotic weld-bead grinding and polishing system [J]. Journal of Mechanical Engineering, 2013, 49(20): 42-48. (in Chinese)

[4]
AOKI Y, GOFORTH H, SRIVATSAN R A, et al. PointNet LK: Robust & efficient point cloud registration using pointnet [C]//Proceedings of 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Los Alamitos, USA: IEEE, 2019: 7163-7172.
[5]
LU W X, WAN G W, ZHOU Y, et al. Deep VCP: An end-to-end deep neural network for point cloud registration [C]//Proceedings of 2019 IEEE/CVF International Conference on Computer Vision. Seoul, Korea (South): IEEE, 2019: 12-21.
[6]
WANG Y, SOLOMON J. Deep closest point: Learning representations for point cloud registration [C]//Proceedings of 2019 IEEE/CVF International Conference on Computer Vision. Seoul, Korea (South): IEEE, 2019: 3523-3532.
[7]
CHARLES R Q, SU H, KAICHUN M, et al. PointNet: Deep learning on point sets for 3D classification and segmentation [C]//Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, USA: IEEE: 2017: 77-85.
[8]
LI J H, ZHANG C H, XU Z Y, et al. Iterative distance-aware similarity matrix convolution with mutual-supervised point elimination for efficient point cloud registration [C]//Proceedings of the 16th European Conference on Computer Vision. Glasgow, UK: Springer, 2020: 378-394.
[9]
YEW Z J, LEE G H. RPM-Net: Robust point matching using learned features [C]//Proceedings of 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Seattle, USA: IEEE, 2020: 11824-11833.
[10]

BESL P J, MCKAY N D. A method for registration of 3-D shapes [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1992, 14(2): 239-256.

[11]

YANG J L, LI H D, CAMPBELL D, et al. Go-ICP: A globally optimal solution to 3D ICP point-set registration [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2016, 38(11): 2241-2254.

[12]
YANG J L, LI H D, and JIA Y D. Go-ICP: Solving 3D registration efficiently and globally optimally [C]//Proceedings of the 2013 IEEE International Conference on Computer Vision. Los Alamitos: IEEE, 2013: 1457-1464.
[13]
FROME A, HUBER D, KOLLURI R, et al. Recognizing objects in range data using regional point descriptors [C]//Proceedings of the 8th European Conference on Computer Vision. Prague, Czech Republic: Springer, 2004: 224-237.
[14]
RUSU R B, BLODOW N, MARTON Z C, et al. Aligning point cloud views using persistent feature histograms [C]//Proceedings of 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems. Nice, France: IEEE, 2008: 3384-3391.
[15]
RUSU R B, BLODOW N, BEETZ M. Fast point feature histograms (FPFH) for 3D registration [C]//Proceedings of 2009 IEEE International Conference on Robotics and Automation. Kobe, Japan: IEEE, 2009: 3212-3217.
[16]

YANG Y, Chen W L, Wang M Y, et al. Color point cloud registration based on supervoxel correspondence [J]. IEEE Access, 2020, 8: 7362-7372

[17]

JUNG S, SONG S, CHANG M, et al. . Range image registration based on 2D synthetic images [J]. Computer-Aided Design, 2018. 94: 16-27.

[18]
ZENG A, SONG S R, NIEBNER M, et al. 2017. 3D match: Learning local geometric descriptors from RGB-D reconstructions [C]//Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, USA: IEEE. 2017: 199-208.
[19]

E A, SEKIKAWA Y, ISHIKAWA K, et al. CorsNet: 3D point cloud registration by deep neural network [J]. IEEE Robotics and Automation Letters, 2020 (3): 3960-3966.

[20]
KENNEDY J, EBERHART R. Particle swarm optimization [C]//Proceedings of the ICNN'95- International Conference on Neural Networks. Perth, Australia: IEEE, 1995: 1942-1948.
[21]
GE Y Q, WANG B Y, NIE J H, et al. A point cloud registration method combining enhanced particle swarm optimization and iterative closest point method [C]//Proceedings of 2016 Chinese Control and Decision Conference. Yinchuan, China: IEEE, 2016: 2810-2815.
[22]

HAN X Q, ZHU Q, DING Y L, et al. Precise registration of scattered cloud data based on the particle swarm optimization [J]. Geomatics and Information Science of Wuhan University, 2014, 39(10): 1214-1220. (in Chinese)

[23]

PAN F, CHEN J, GAN M G, et al. Model analysis of particle swarm optimizer [J]. Acta Automatica Sinica, 2006, 32(3): 368-377. (in Chinese)

[24]

ZHANG D X, GUAN Z H, LIU X Z. Adaptive particle swarm optimization algorithm with dynamically changing inertia weight [J]. Control and Decision, 2008, 23(11): 1253-1257. (in Chinese)

[25]

WANG J W, WANG D W. Particle swarm optimization algorithm with gradient acceleration [J]. Control and Decision, 2004(11): 1298-1300, 1304. (in Chinese)

Journal of Tsinghua University (Science and Technology)
Pages 1367-1379

{{item.num}}

Comments on this article

Go to comment

< Back to all reports

Review Status: {{reviewData.commendedNum}} Commended , {{reviewData.revisionRequiredNum}} Revision Required , {{reviewData.notCommendedNum}} Not Commended Under Peer Review

Review Comment

Close
Close
Cite this article:
LIU H, GUO Z, NIE H, et al. Online intelligent positioning of parts based on adaptive inspired particle swarm optimization. Journal of Tsinghua University (Science and Technology), 2024, 64(8): 1367-1379. https://doi.org/10.16511/j.cnki.qhdxxb.2024.27.016

307

Views

8

Downloads

0

Crossref

0

Scopus

0

CSCD

Altmetrics

Received: 13 November 2023
Published: 15 August 2024
© Journal of Tsinghua University (Science and Technology). All rights reserved.