Blog posts

2024

Aligning LLMs to User Preference

8 minute read

Published:

Table of Contents:

  • Why LLM Alignment
  • Reinforcement learning from human feedback
  • Direct Preference Optimization
  • Kahneman-Tversky Optimization
  • Survey studies
  • Miscellaneous