Reinforcement Learning Foundations for Deep Research Systems: A Survey Paper • 2509.06733 • Published Sep 8 • 31
Bootstrapping Language Models with DPO Implicit Rewards Paper • 2406.09760 • Published Jun 14, 2024 • 40
Aligning Crowd Feedback via Distributional Preference Reward Modeling Paper • 2402.09764 • Published Feb 15, 2024 • 1