Qidong Su 蘇起冬
My first name is pronounced as /tɕʰi tʊŋ/.
I am a Computer Science PhD student in University of Toronto, advised by Gennady Pekhimenko. I got my bachelor degree from Shanghai Jiao Tong University (ACM Class).
My research focuses on accelerating programs on modern hardware. Currently I am optimizing inference speed of large-scale models.
I'm happy to discuss topics including (but not limited to) machine learning systems, compiler designs, parallel programming, etc.
An amateur in linguistics (phonetics, Chinese dialects, Japanese).
Publications
-
[MLSys 2025] Seesaw: High-throughput LLM Inference via Model Re-sharding
-
[EuroSys 2025] Mist: Efficient Distributed Training of Large Language Models via Memory-Parallelism Co-Optimization
-
[PACT 2024] BOOM: Use your Desktop to Accurately Predict the Performance of Large Deep Neural Networks
-
[arXiv] APPL: A Prompt Programming Language for Harmonious Integration of Programs and Large Language Model Prompts
-
[COLM 2024] A Survey on Deep Learning for Theorem Proving
-
[arXiv] The Synergy of Speculative Decoding and Batching in Serving Large Language Models
-
[APLAS 2023] TorchProbe: Fuzzing Dynamic Deep Learning Compilers
-
[GNNSys 2021] Adaptive Load Balancing for Parallel GNN Training
-
[IA3 2021 @SC] DistDGL: Distributed Graph Neural Network Training for Billion-Scale Graphs
Contact
- Email: qdsu ät cs.toronto.edu
Last modified by Qidong Su. (Feb.11 2025)