I am a fourth-year Ph.D. student in Computer Science Department at University of California, Los Angeles. I am generally interested in cloud computing and ML systems. I am co-advised by Professor Harry Xu and Professor Miryung Kim. I received the Amazon & UCLA Science Hub Fellowship in 2022.
Prior to graduate school, I earned my B.E. in Computer Science from Tsinghua University in 2019, and I was a research intern in PACMAN group and Storage research group. I also worked with Professor Umut Acar on scheduling algorithms for multithreaded parallel computing in 2018.
I work in the area of System for Cloud Computing and Machine Learning. My research spans machine learning algorithms, distributed systems, and operating systems.
In particular, I have worked on projects that aim to lower the barriers of training large-scale ML models on cloud and improve datacenter resource utilization with OS and runtime innovations. The underlying theme behind my research thrust is to design seamless, tightly integrated, vertical systems: (1) to co-design training algorithms and serverless computing platforms for distributed graph neural network training, (2) to co-design runtimes and OS kernel for adaptive and semantic-aware remote memory datapath, and (3) to re-design OS for isolated and efficient remote memory management.
Hermit: Low-Latency, High-Throughput, and Transparent Remote Memory via Feedback-Directed Asynchrony
Yifan Qiao, Chenxi Wang, Zhenyuan Ruan, Adam Belay, Qingda Lu, Yiying Zhang, Miryung Kim, and Guoqing Harry Xu.
The USENIX Symposium on Networked Systems Design and Implementation (NSDI), 2023 (to appear)
Canvas: Isolated and Adaptive Swapping for Multi-Applications on Remote Memory
Chenxi Wang*, Yifan Qiao*, Haoran Ma, Shi Liu, Yiying Zhang, Wenguang Chen, Ravi Netravali, Miryung Kim, Guoqing Harry Xu. (* contributed equally)
The USENIX Symposium on Networked Systems Design and Implementation (NSDI), 2023 (to appear)
MemLiner: Lining up Tracing and Application for a Far-Memory-Friendly Runtime
Chenxi Wang*, Haoran Ma*, Shi Liu, Yifan Qiao, Jonathan Eyolfson, Christian Navasca, Shan Lu, Guoqing Harry Xu. (* contributed equally)
The USENIX Symposium on Operating Systems Design and Implementation (OSDI), 2022
(Awarded Jay Lepreau Best Paper)
[code]
Mako: A Low-Pause, High-Throughput Evacuating Collector for Memory-Disaggregated Datacenters
Haoran Ma, Shi Liu, Chenxi Wang, Yifan Qiao, Michael D. Bond, Stephen M. Blackburn, Miryung Kim, Guoqing Harry Xu.
The ACM SIGPLAN Conference on Programming Language Design and Implementation (PLDI), 2022
[code]
Bamboo: Making Preemptible Instances Resilient for Affordable Training of Large DNNs
John Thorpe*, Pengzhan Zhao*, Jonathan Eyolfson, Yifan Qiao, Zhihao Jia, Minjia Zhang, Ravi Netravali, Guoqing Harry Xu. (* contributed equally)
The USENIX Symposium on Networked Systems Design and Implementation (NSDI), 2023 (to appear)
[full version] [code]Dorylus: Affordable, Scalable, and Accurate GNN Training over Billion-Edge Graphs
John Thorpe*, Yifan Qiao*, Jonathan Eyolfson, Shen Teng, Guanzhou Hu, Zhihao Jia, Jinliang Wei, Keval Vora, Ravi Netravali, Miryung Kim, and Guoqing Harry Xu. (* contributed equally)
The USENIX Symposium on Operating Systems Design and Implementation (OSDI), 2021
[full version] [code]
Algorithm-Directed Crash Consistence in Non-Volatile Memory for HPC
Shuo Yang, Kai Wu, Yifan Qiao, Dong Li, Jidong Zhai.
IEEE International Conference on Cluster Computing (CLUSTER), 2017
WORDS 2022, Session Chair