PERL: Parameter Efficient Reinforcement Learning from Human Feedback Paper • 2403.10704 • Published Mar 15, 2024 • 57
WARM: On the Benefits of Weight Averaged Reward Models Paper • 2401.12187 • Published Jan 22, 2024 • 18
RewardBench: Evaluating Reward Models for Language Modeling Paper • 2403.13787 • Published Mar 20, 2024 • 21
DreamReward: Text-to-3D Generation with Human Preference Paper • 2403.14613 • Published Mar 21, 2024 • 35
Direct Preference Optimization: Your Language Model is Secretly a Reward Model Paper • 2305.18290 • Published May 29, 2023 • 50
DRLC: Reinforcement Learning with Dense Rewards from LLM Critic Paper • 2401.07382 • Published Jan 14, 2024 • 2