The N+ Implementation Details of RLHF with PPO: A Case Study on TL;DR Summarization Paper • 2403.17031 • Published Mar 24, 2024 • 3
Asynchronous RLHF: Faster and More Efficient Off-Policy RL for Language Models Paper • 2410.18252 • Published Oct 23, 2024 • 5
TÜLU 3: Pushing Frontiers in Open Language Model Post-Training Paper • 2411.15124 • Published Nov 22, 2024 • 58
If You Can't Use Them, Recycle Them: Optimizing Merging at Scale Mitigates Performance Tradeoffs Paper • 2412.04144 • Published Dec 5, 2024 • 4
RLHF Can Speak Many Languages: Unlocking Multilingual Preference Optimization for LLMs Paper • 2407.02552 • Published Jul 2, 2024 • 4
Pushing Mixture of Experts to the Limit: Extremely Parameter Efficient MoE for Instruction Tuning Paper • 2309.05444 • Published Sep 11, 2023 • 1
Back to Basics: Revisiting REINFORCE Style Optimization for Learning from Human Feedback in LLMs Paper • 2402.14740 • Published Feb 22, 2024 • 12
cleanrl/EleutherAI_pythia-1b-deduped__reward__tldr Text Classification • Updated May 15, 2024 • 1.96k