Yangjun Ruan

Ph.D. Candidate, Computer Science
Machine Learning Group
University of Toronto & Vector Institute

Email: yjruan [at] cs [dot] toronto [dot] edu



I am a Ph.D. student in Computer Science at University of Toronto, where I am fortunate to be advised by Chris Maddison and Jimmy Ba. Currently, I am also a visiting scholar at Stanford University, hosted by Tatsunori Hashimoto.

Previously, I was a student researcher at Google Research and a research intern at Microsoft Research. In summer 2019, I was a visiting student at UCLA, where I worked with Cho-Jui Hsieh. I obtained my Bachelor degree in Information Engineering from Zhejiang University.

I am on the industrial job market now.

Research


My research focuses on the new scaling paradigms of language models and agents in data-constrained scenarios. I am generally interested in synthetic data, scalable evaluation and alignment, and agents.


Selected Publications [Full List]


* below denotes equal contribution

  1. Reasoning to Learn from Latent Thoughts
    Yangjun Ruan, Neil Band, Chris J Maddison, and Tatsunori Hashimoto
    arXiv preprint arXiv:2503.18866, 2025
  2. Putting It All into Context: Simplifying Agents with LCLMs
    Mingjian Jiang, Yangjun Ruan, Luis Lastras, Pavan Kapanipathi, and Tatsunori Hashimoto
    arXiv preprint arXiv:2505.08120, 2025
  3. Observational Scaling Laws and the Predictability of Language Model Performance
    Yangjun Ruan, Chris J Maddison, and Tatsunori Hashimoto
    In Advances in Neural Information Processing Systems (NeurIPS), 2024 [Spotlight]
  4. Identifying the Risks of LM Agents with an LM-Emulated Sandbox
    Yangjun Ruan*, Honghua Dong*, Andrew Wang, Silviu Pitis, Yongchao Zhou, Jimmy Ba, Yann Dubois, Chris J. Maddison, and Tatsunori Hashimoto
    In International Conference on Learning Representations (ICLR), 2024 [Spotlight]
  5. Optimal Representations for Covariate Shift
    Yangjun Ruan*, Yann Dubois*, and Chris J Maddison
    In International Conference on Learning Representations (ICLR), 2022
  6. Improving Lossless Compression Rates via Monte Carlo Bits-Back Coding
    Yangjun Ruan*, Karen Ullrich*, Daniel Severo*, James Townsend, Ashish Khisti, Arnaud Doucet, Alireza Makhzani, and Chris J Maddison
    In International Conference on Machine Learning (ICML), 2021 [Oral]


Selected Awards & Honors


  • Ontario Graduate Scholarship, 2023
  • DiDi Gruduate Student Award, 2021
  • CHU Kochen Scholarship (highest honor at Zhejiang University), 2019.
  • Cross-disciplinary Scholars in Science and Technology (CSST), UCLA, 2019.
  • National Scholarship (top 1.5%), 2017, 2018, 2019.
  • Meritorious Winner, Interdisciplinary Contest in Modeling (ICM), 2018.