Aligning LLMs to User Preference 8 minute read Published: April 27, 2024Table of Contents:Why LLM AlignmentReinforcement learning from human feedbackDirect Preference OptimizationKahneman-Tversky OptimizationSurvey studiesMiscellaneous