Photo of Yuntian Deng

Yuntian Deng

Assistant Professor, UWaterloo
Associate, Harvard SEAS
Faculty Affiliate, Vector Institute
PhD in CS, Harvard
[CV] [Google Scholar] [GitHub] [X/Twitter] [LinkedIn] [Semantic Scholar]

My research focuses on understanding and improving how language models reason, with a particular emphasis on internalizing explicit reasoning processes into implicit computation. I also study real-world LLM usage at scale through large-scale conversation datasets and build open tools and demos that make our research accessible.

Our work on WildChat has been featured in the Washington Post and is used in OpenAI's o1 and Anthropic's Claude 3 for safety evaluation. Our open-source toolkit OpenNMT has been widely adopted in both industry and academia. I received my PhD from Harvard University, where I was advised by Prof. Alexander Rush and Prof. Stuart Shieber. I did a postdoc under the supervision of Prof. Yejin Choi.


Research Themes

Neural World Models

Revamping how machines interact with humans using generative AI instead of rigid menus and rules. We build neural models that simulate entire computing environments end-to-end.

NeuralOS
Program as Weights

A new programming paradigm that replaces symbolic code for fuzzy functions with compiled neural programs. Towards programming without code as an intermediary.

programasweights.com
Implicit Reasoning

Can language models learn to reason without explicitly articulating every step? We develop methods to internalize chain-of-thought reasoning into a model's hidden computation.

Internalize CoT · Implicit CoT · ICLR 2026 Workshop
LLM Usage at Scale

How do people actually use large language models? We collect and analyze millions of real-world conversations to understand usage patterns, safety, and emergent behaviors.

WildChat · WildVis · Aggregative QA

In the Media


News


Current Research

Figure from NeuralOS: Towards Simulating Operating Systems via Neural Generative Models
NeuralOS: Towards Simulating Operating Systems via Neural Generative Models
Luke Rivard, Sun Sun, Hongyu Guo, Wenhu Chen, Yuntian Deng.
ICLR 2026
Figure from Program as Weights
Program as Weights
Wentao Zhang*, Liliana Hotsko*, Woojeong Kim*, Pengyu Nie, Stuart Shieber, Yuntian Deng.
Coming soon
Figure from From Explicit CoT to Implicit CoT: Learning to Internalize CoT Step by Step
From Explicit CoT to Implicit CoT: Learning to Internalize CoT Step by Step
Yuntian Deng, Yejin Choi, Stuart Shieber.
arXiv 2024
Figure from Implicit Chain of Thought Reasoning via Knowledge Distillation
Implicit Chain of Thought Reasoning via Knowledge Distillation
Yuntian Deng, Kiran Prasad, Roland Fernandez, Paul Smolensky, Vishrav Chaudhary, Stuart Shieber.
arXiv 2023
Figure from WildChat: 1M ChatGPT Interaction Logs in the Wild
WildChat: 1M ChatGPT Interaction Logs in the Wild
Wenting Zhao, Xiang Ren, Jack Hessel, Claire Cardie, Yejin Choi, Yuntian Deng.
ICLR 2024 (Spotlight)
Featured in the Washington Post
Used in OpenAI's o1 for safety evaluation
Used in Anthropic's Claude 3 for evaluating refusals

Other Selected Works

For all my papers, visit the publications page or my Google Scholar.

Figure from From Chat Logs to Collective Insights: Aggregative Question Answering
From Chat Logs to Collective Insights: Aggregative Question Answering
Wentao Zhang, Woojeong Kim, Yuntian Deng.
EMNLP 2025 (Oral)
Figure from Interactive Training: Feedback-Driven Neural Network Optimization
Interactive Training: Feedback-Driven Neural Network Optimization
Wentao Zhang, Yang Young Lu, Yuntian Deng.
EMNLP 2025 Demo
Figure from WildVis: Open Source Visualizer for Million-Scale Chat Logs in the Wild
WildVis: Open Source Visualizer for Million-Scale Chat Logs in the Wild
Yuntian Deng, Wenting Zhao, Jack Hessel, Xiang Ren, Claire Cardie, Yejin Choi.
EMNLP 2024 Demo
Figure from Tree Prompting: Efficient Task Adaptation without Fine-Tuning
Tree Prompting: Efficient Task Adaptation without Fine-Tuning
John Xavier Morris*, Chandan Singh*, Alexander M. Rush, Jianfeng Gao, Yuntian Deng.
EMNLP 2023
Figure from Markup-to-Image Diffusion Models with Scheduled Sampling
Markup-to-Image Diffusion Models with Scheduled Sampling
Yuntian Deng, Noriyuki Kojima, Alexander M. Rush.
ICLR 2023
Figure from Model Criticism for Long-Form Text Generation
Model Criticism for Long-Form Text Generation
Yuntian Deng, Volodymyr Kuleshov, Alexander M Rush.
EMNLP 2022
Figure from Cascaded Text Generation with Markov Transformers
Cascaded Text Generation with Markov Transformers
Yuntian Deng, Alexander M. Rush.
NeurIPS 2020
Figure from Residual Energy-Based Models for Text Generation
Residual Energy-Based Models for Text Generation
Yuntian Deng, Anton Bakhtin, Myle Ott, Arthur Szlam, Marc'Aurelio Ranzato.
ICLR 2020
Referenced by Meta's Llama 2
Referenced by the diffusion paper (DDPM)
Figure from Bottom-Up Abstractive Summarization
Bottom-Up Abstractive Summarization
Sebastian Gehrmann, Yuntian Deng, Alexander Rush.
EMNLP 2018
Figure from Latent Alignment and Variational Attention
Latent Alignment and Variational Attention
Yuntian Deng*, Yoon Kim*, Justin Chiu, Demi Guo, Alexander M. Rush.
NeurIPS 2018
Figure from Image-to-Markup Generation with Coarse-to-Fine Attention
Image-to-Markup Generation with Coarse-to-Fine Attention
Yuntian Deng, Anssi Kanervisto, Jeffrey Ling, and Alexander M. Rush.
ICML 2017
Figure from Neural Linguistic Steganography
Neural Linguistic Steganography
Zachary Ziegler*, Yuntian Deng*, Alexander Rush.
EMNLP 2019 (Oral)
Figure from OpenNMT: Open-Source Toolkit for Neural Machine Translation
OpenNMT: Open-Source Toolkit for Neural Machine Translation
Guillaume Klein, Yoon Kim, Yuntian Deng, Jean Senellart, Alexander M. Rush.
ACL 2017 Demo (Best Demo Runner-up)

Service


Prospective Students

I am not actively recruiting new students at this time. If you are genuinely interested in my research, I encourage you to read my recent papers and reach out with specific questions about the work. Due to high volume, I am unable to respond to generic inquiries.