I am now at a stealth startup.
I was at Google in the N2Formal team led by Christian Szegedy.
I did Postdoc at Stanford with Percy Liang
and Jay McClelland.
During my PhD at U of Toronto, I was advised by Roger Grosse and Jimmy Ba.
My primary research interest is building machines that can reason.
I have chosen mathematics as a starting point to study reasoning, with the aim of creating an automated mathematician.
July. 2023 |
A New York Times story covering our work. |
Feb. 2023 |
Our work is featured in Quanta Magazine! |
Jan. 2023 | |
Nov. 2022 |
Releasing Draft, Sketch, and Prove: Autoformalize the entire natural language proofs [arxiv]! I gave a talk on autoformalization at FLAIM conference. I gave a guest lecture on autoformalization at UIUC proof automation class. |
Sept. 2022 |
8 papers accepted to NeurIPS 2022. Our length generalization paper is accepted as an Oral Presentation at NeurIPS 2022. We are organizing the second MATHAI workshop at NeurIPS 2022. I gave a talk at AITP 2022. |
June 2022 |
Releasing Minerva: a language model that solves MATH with 51% acc, which was predicted to happen in 2025! See [arXiv][Google AI Blog][Sample Explorer]. Sharing a systematic study on synthetic pre-training [arXiv]. Understanding pre-training via synthetic tasks! I gave a talk at the University of Cambridge [Link]. I gave talk at UC Berkeley Center for Human-Compatible AI (CHAI). I gave a talk at Covariant.ai. |
May 2022 |
We used LLMs to turn natural language mathematics into formal specifications [arXiv], and achieved SOTA on miniF2F. See media coverage on NewScientist! We released Thor [arXiv]. Integrate symbolic tools to neural theorem provers for premise selection! We released a stronger version of Subgoal search. Introduce Adaptive Subgoal Search (AdaSubS) [arXiv]: improve search with transformers by variable planning horizons.
|
March 2022 |
We released STaR [arXiv]. Bootstrapping Reasoning with Reasoning! We released Block-Recurrent Transformer [arXiv]. Recurrence is coming back! Gave a talk at the University of Oxford. Gave a talk at Harvard University. |
Jan 2022 |
Memorizing Transformers accepted as a spotlight presentation at ICLR 2022. Three papers accepted to ICLR 2022. |
Dec 2021 |
Subgoal search algorithm accepted to NeurIPS 2021. |
Co-organized the MATHAI4ED workshop at NeurIPS 2021: Math AI for education: Bridging the gap between research and smart education. | |
Aug 2021 |
Led the Reasoning section in the Foundation Model white paper. |
Jul 2021 |
Two posters in ICML 2021. |
Apr 2021 |
Two posters in ICLR 2021. |
Draft, Sketch, and Prove: Guiding Formal Theorem Provers with Informal Proofs
Albert Qiaochu Jiang*, Sean Welleck*, Jin Peng Zhou*, Timothee Lacroix, Jiacheng Liu, Wenda Li,
Mateja Jamnik, Guillaume Lample, Yuhuai Wu
The 10th International Conference on Learning Representations, 2023.
PDF
Minerva: Solving Quantitative Reasoning Problems with Language Models
Aitor Lewkowycz, Anders Andreassen, David Dohan, Ethan Dyer, Henryk Michalewski,
Vinay Ramasesh, Ambrose Slone, Cem Anil, Imanol Schlag, Theo Gutman-Solo,
Yuhuai Wu, Behnam Neyshabur, Guy Gur-Ari, Vedant Misra
NeurIPS, 2022.
PDF Google AI Blog
Exploring Length Generalization in Large Language Models
Cem Anil, Yuhuai Wu, Anders Andreassen, Aitor Lewkowycz, Vedant Misra,
Vinay Ramasesh, Ambrose Slone, Guy Gur-Ari, Ethan Dyer, Behnam Neyshabur
The 36th Conference on Neural Information Processing Systems, 2022.
PDF
Insights into Pre-training via Simpler Synthetic Tasks
Yuhuai Wu*,
Felix Li*,
Percy Liang
NeurIPS, 2022.
PDF
Autoformalization with Large Language Models
Yuhuai Wu*,
Albert Q. Jiang,
Wenda Li,
Markus Rabe,
Charles Staats,
Mateja Jamnik,
Christian Szegedy
NeurIPS, 2022.
PDF Interview with NewScientist
STaR: Bootstrapping Reasoning With Reasoning
Eric Zelikman*, Yuhuai Wu*, Noah D. Goodman
arXiv, 2022.
PDF
Block-Recurrent Transformers.
DeLesley Hutchins*, Imanol Schlag*, Yuhuai Wu, Ethan Dyer, Behnam Neyshabur
arXiv, 2022.
PDF
Memorizing Transformers.
Yuhuai Wu,
Markus Rabe,
DeLesley Hutchins,
Christian Szegedy
The 10th International Conference on Learning Representations, 2022.
PDF #4 on HackerNews
Proof Artifact Co-training for Theorem Proving with Language Model.
Jesse Michael Han,
Jason Rute, Yuhuai Wu, Edward W. Ayers, Stanislas Polu
The 10th International Conference on Learning Representations, 2022.
PDF
Subgoal Search For Complex Reasoning Tasks.
Konrad Czechowski, Tomasz Odrzygozdz, Marek Zbysinski, Michal Zawalski,
Krzysztof Olejnik,Yuhuai Wu,
Lukasz Kucinski, Piotr Milos
The 35th Conference on Neural Information Processing Systems, 2021.
PDF
LIME: Learning Inductive Bias for Primitives of Mathematical Reasoning.
Yuhuai Wu,
Markus Rabe,
Wenda Li,
Jimmy Ba,
Roger Grosse,
Christian Szegedy
The 38th International Conference on Machine Learning, 2021.
PDF
INT: An Inequality Benchmark for Evaluating Generalization in Theorem Proving.
Yuhuai Wu*,
Albert Q. Jiang*,
Jimmy Ba,
Roger Grosse
The 9th International Conference on Learning Representations, 2021.
PDF
Modelling High-Level Mathematical Reasoning in Mechanised Declarative Proofs.
Wenda Li, Lei Yu, Yuhuai Wu, Lawrence C. Paulson
The 9th International Conference on Learning Representations, 2021.
PDF
Learning to Give Checkable Answers with Prover-Verifier Games.
Cem Anil,
Guodong Zhang,
Yuhuai Wu,
Roger Grosse
2021
PDF
On the Opportunities and Risks of Foundation Models.
Rishi Bommasani, Drew A. Hudson, Percy Liang et. al.
Options as REsponses: Grounding Behavioural Hierarchies in Multi-agent Reinforcement Learning.
Yuhuai Wu*, Alexander Sasha Vezhnevets*, Maria Eckstein, Remi Leblond, Joel Z. Leibo.
The 37th International Conference on Machine Learning, 2020.
PDF
Grandmaster Level in StarCraft II using Multi-gent Reinforcement Learning.
Vinyals, O., Babuschkin, I., Czarnecki, W.M. et al.
Nature, 2019.
PDF
Understanding Short-Horizon Bias in Stochastic Meta-Optimization.
Yuhuai Wu*,
Mengye Ren*,
Renjie Liao,
Roger Grosse
The 6th International Conference on Learning Representations, 2018.
PDF
Scalable Trust-Region Method for Deep Reinforcement Learning using Kronecker-Factored Approximation.
Yuhuai Wu*, Elman Mansimov*, Shun Liao,
Roger Grosse,
Jimmy Ba
The 31st Annual Conference on Neural Information Processing Systems, 2017
PDF
On the Quantitative Analysis of Decoder-Based Generative Models
Yuhuai Wu,
Yuri Burda,
Ruslan Salakhutdinov,
Roger Grosse
The 5th International Conference on Learning Representations, 2017
PDF