About

My name is Guodong Liu, I am currently a Ph.D. candidate in the Institute of Computing Technology (ICT), Chinese Academy of Sciences (CAS), supervised by Prof. Yungang Bao and Prof. Sa Wang. I received my bachelor degree in Computer Science from University of Chinese Academy of Sciences (UCAS) in 2019. My full CV is available here: [CV]

My research interest lies in the intersection of machine learning and systems. This website is for posting some of my surveys and thoughts on recent research. If you are interested on similar topics, please contact me and have a talk.

I am on the job market for 2025. Please feel free to reach out if you have openings.

Selected Research

Aceso: Efficient Parallel DNN Training through Iterative Bottleneck Alleviation
Guodong Liu, Youshan Miao, Zhiqi Lin, Xiaoxiang Shi, Saeed Maleki, Fan Yang, Yungang Bao, Sa Wang
EuroSys 2024

Aceso is a scalable parallel-mechanism auto-configuring system that operates iteratively. For a given parallel configuration, Aceso identifies a performance bottleneck and then, by summarizing all possible configuration adjustments with their resource consumption changes, infers their performance impacts to the bottleneck and selects one that mitigates the bottleneck. Aceso significantly reduces configuration searching cost by taking the approach of resolving one bottleneck at a time. We implemented and tested Aceso on representative DNN models. Evaluations show that it can scale to 1K-layer models. Compared to state-of-the-art systems, Aceso achieves up to 1.33x throughput improvement with less than 5% of the searching cost.

SEER: A Time Prediction Model for CNNs from GPU Kernel’s View
Guodong Liu, Sa Wang, Yungang Bao
PACT 2021

SEER is an iteration time prediction model for CNNs, targeting on GPU platforms. We propose to categorize convolution kernels into three different types: Compute-bound, DRAM-bound and Under-utilized, then we build performance model for each type respectively. We combined analytical models and learning-based models to make the performance model accurate and in line with GPU execution model. Experimental results show that, our model achieves 14.71% prediction error on convolution kernels and up to 1.79% prediction error for the overall computation time in one iteration of common CNNs. Besides, when used for selecting the best convolution algorithm, our model shows 7.14% lower error rate than cuDNN’s official algorithm picker.

» More publications

Featured Article

March 2022: Interviewed by Microsoft Research Asia:

Arts

July 2022: My painting “World on Silicon” got selected as cover of the journal Science Writing Review:

world on silicon