Journal Home > Volume 36 , Issue 4

It is challenging to model the performance of distributed graph computation. Explicit formulation cannot easily capture the diversified factors and complex interactions in the system. Statistical learning methods require a large number of training samples to generate an accurate prediction model. However, it is time-consuming to run the required graph computation tests to obtain the training samples. In this paper, we propose TransGPerf, a transfer learning based solution that can exploit prior knowledge from a source scenario and utilize a manageable amount of training data for modeling the performance of a target graph computation scenario. Experimental results show that our proposed method is capable of generating accurate models for a wide range of graph computation tasks on PowerGraph and GraphX. It outperforms transfer learning methods proposed for other applications in the literature.

File
jcst-36-4-778-Highlights.pdf (175.7 KB)
Publication history
Copyright
Acknowledgements

Publication history

Received: 02 February 2021
Accepted: 10 July 2021
Published: 05 July 2021
Issue date: July 2021

Copyright

©Institute of Computing Technology, Chinese Academy of Sciences 2021

Acknowledgements

Acknowledgement(s)

We would like to thank Deepayan Chakrabarti for the graph statistics package NetMine and Himchan Park for discussions about graph characteristics, which help a lot for our graph feature extraction.

Return