Bottom-up Policy Optimization: Your Language Model Policy Secretly Contains Internal Policies Paper • 2512.19673 • Published 12 days ago • 60
On the Interplay of Pre-Training, Mid-Training, and RL on Reasoning Language Models Paper • 2512.07783 • Published 26 days ago • 36
ASPO: Asymmetric Importance Sampling Policy Optimization Paper • 2510.06062 • Published Oct 7, 2025 • 13
Attention as a Compass: Efficient Exploration for Process-Supervised RL in Reasoning Models Paper • 2509.26628 • Published Sep 30, 2025 • 16
Random Policy Valuation is Enough for LLM Reasoning with Verifiable Rewards Paper • 2509.24981 • Published Sep 29, 2025 • 29
A Survey of Reinforcement Learning for Large Reasoning Models Paper • 2509.08827 • Published Sep 10, 2025 • 190
Part I: Tricks or Traps? A Deep Dive into RL for LLM Reasoning Paper • 2508.08221 • Published Aug 11, 2025 • 50
Stabilizing Knowledge, Promoting Reasoning: Dual-Token Constraints for RLVR Paper • 2507.15778 • Published Jul 21, 2025 • 20
Scaling Image and Video Generation via Test-Time Evolutionary Search Paper • 2505.17618 • Published May 23, 2025 • 41
GenPRM: Scaling Test-Time Compute of Process Reward Models via Generative Reasoning Paper • 2504.00891 • Published Apr 1, 2025 • 14
GenPRM Collection A collection of GenPRM. Project page: https://ryanliu112.github.io/GenPRM • 6 items • Updated Apr 6, 2025 • 5
CodeI/O Collection Collection for CodeI/O @ https://codei-o.github.io/ • 16 items • Updated May 6, 2025 • 7
VersaPRM Collection Collection of VersaPRMs using various training configurations • 8 items • Updated Feb 8, 2025 • 1
Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling Paper • 2502.06703 • Published Feb 10, 2025 • 153