Structured Metacognition via 2-Level GRPO-Embedded Monte Carlo Tree Search for Decoding Policy Optimization

Niel Ok
Stanford University ยท May 2025
Code

TL;DR: We introduce a test-time inference framework, GRPO-embedded Monte Carlo Tree Search (GRPO-MCTS), that enables text diffusion models to optimize how they decode, not just what they decode. Rather than relying on static decoding strategies, our approach performs structured search over blockwise inference schedules customized for each input prompt.

Each decoding policy is a structured program specifying temperature, remasking strategy, step allocation, and compute distribution. GRPO-MCTS explores this policy space using Monte Carlo Tree Search, while applying Group-Relative Policy Optimization (GRPO) at each node to learn reward-aligned sampling behavior. This two-level structure, search over decoding blocks and local optimization over rollouts, forms a dynamic system of structured metacognition. Scaling is controlled by the breadth of tree search and the precision of GRPO updates.

This work is the first to frame inference-time decoding as a real-time optimization problem over compositional strategies, demonstrating that language models can adapt how they think on a per-prompt basis.

We gratefully acknowledge compute support provided by researchers at DeepMind in preparation for DeepMind's Gemini Diffusion release.

BibTeX

@misc{ok2025grpomcts,
  title        = {Structured Metacognition via 2-Level GRPO-Embedded Monte Carlo Tree Search for Decoding Policy Optimization},
  author       = {Niel Ok},
  year         = {2025},
  month        = {May},
  url          = {https://nielok.github.io/LLaDA_TTC/}
}