AlphaApollo: Orchestrating Foundation Models and Professional Tools into a Self-Evolving System for Deep Agentic Reasoning Paper • 2510.06261 • Published Oct 5, 2025 • 6
Reasoned Safety Alignment: Ensuring Jailbreak Defense via Answer-Then-Check Paper • 2509.11629 • Published Sep 15, 2025 • 1
Envisioning Outlier Exposure by Large Language Models for Out-of-Distribution Detection Paper • 2406.00806 • Published Jun 2, 2024
Co-rewarding Collection Co-rewarding is a novel self-supervised RL framework that improves training stability by seeking complementary supervision from another views. • 75 items • Updated Dec 21, 2025 • 1
Co-rewarding Collection Co-rewarding is a novel self-supervised RL framework that improves training stability by seeking complementary supervision from another views. • 75 items • Updated Dec 21, 2025 • 1