avatar

Sizhuang He

Second-Year Ph.D. Student in Computer Science
Yale University
sizhuang.he (at) yale.edu

Loading...
Citations
Loading...
H-Index
Last updated: Loading...

About Me

I’m a second-year Computer Science Ph.D. student at Yale University, advised by Dr. David van Dijk. Previously, I completed my Bachelor’s degree at University of Michigan, Ann Arbor, majoring in Honors Mathematics and minoring in Computer Science.

My research interests include generative modeling, LLM agents, and post-training. I have worked on discrete diffusion models that unify diffusion and autoregressive generation (CaDDi, NeurIPS 2025) and learn distributions over finite symmetric groups (Soft-Rank Diffusion). I also build LLM-based multi-agent systems for automating scientific data curation at scale. Currently, I’m exploring dense, verifiable reward design for training LLM agents via reinforcement learning.

News

Selected Publications

  1. Soft-Rank Diffusion
    We propose Soft-Rank Diffusion, a discrete diffusion framework for learning distributions over permutations on $S_n$, using continuous relaxed ranks and contextualized generalized Plackett–Luce (cGPL) denoisers. Soft-Rank Diffusion consistently outperforms prior baselines, with especially strong gains on long sequences. — In Review

  2. Cell2Sentence
    C2S-Scale scales this framework to 27 billion parameters trained on a billion-token multimodal corpus—achieving state-of-the-art predictive and generative performance for complex, multicellular analyses. Visit the project page. Read more about this work in our blog post and another blog post. — In Review

  3. CaDDI
    We introduce a novel approach to discrete diffusion models that conditions on the entire generative trajectory, thereby lifting the Markov constraint and allowing the model to revisit and improve past states. CaDDi treats standard causal language models as a special case and permits the direct reuse of pretrained LLM weights with no architectural changes. — NeurIPS 2025

  4. Intelligence at the Edge of Chaos
    By training LLMs on elementary cellular automata rules of varying complexity, we pinpoint a ‘sweet spot’ of data complexity that maximizes downstream predictive and reasoning abilities. Our findings suggest that exposing models to appropriately complex patterns is key to unlocking emergent intelligence. — ICLR 2025

Services

Journal Reviewer

Conference Reviewer


Powered by Jekyll and Minimal Light theme.