The Alignment Waltz: Jointly Training Agents to Collaborate for Safety Paper • 2510.08240 • Published 24 days ago • 40
Learning to Optimize Multi-Objective Alignment Through Dynamic Reward Weighting Paper • 2509.11452 • Published Sep 14 • 13
Benchmarking Optimizers for Large Language Model Pretraining Paper • 2509.01440 • Published Sep 1 • 24
Jointly Reinforcing Diversity and Quality in Language Model Generations Paper • 2509.02534 • Published Sep 2 • 24
Certified Mitigation of Worst-Case LLM Copyright Infringement Paper • 2504.16046 • Published Apr 22 • 13
RATIONALYST: Pre-training Process-Supervision for Improving Reasoning Paper • 2410.01044 • Published Oct 1, 2024 • 36
Benchmarking Language Model Creativity: A Case Study on Code Generation Paper • 2407.09007 • Published Jul 12, 2024 • 4