Dr.LLM: Dynamic Layer Routing for LLMs

Abstract

Dr.LLM improves reasoning with less computation. Lightweight routers at each layer decide whether to skip, execute, or repeat transformer blocks. Monte Carlo tree search trains the routers to find optimal paths. The approach is retrofittable and generalises to new tasks.

Publication
International Conference on Learning Representations 2026