2024/11 45

RS-DPO: A Hybrid Rejection Sampling and Direct PreferenceOptimization Method for Alignment of Large Language Models

https://arxiv.org/pdf/2402.10038Reinforcement learning from human feedback (RLHF) has been extensively employed to align large language models with user intent. However, proximal policy optimization (PPO) based RLHF is occasionally unstable requiring significant hyperparameter finetuning, and computationally expensive to maximize the estimated reward during alignment. Recently, direct preference..

카테고리 없음 2024.11.01

Self-Play Preference Optimization for Language Model Alignment 논문리뷰

https://arxiv.org/pdf/2405.00675Standard reinforcement learning from human feedback (RLHF) approaches relying on parametric models like the Bradley-Terry model fall short in capturing the intransitivity and irrationality in human preferences. Recent advancements suggest that directly working with preference probabilities can yield a more accurate reflection of human preferences, enabling more fl..

카테고리 없음 2024.11.01