https://www.youtube.com/watch?v=vuWbJlBePPA&list=PLgKuh-lKre12qVTl88k2n2N37tT-BpmHT&index=3

Background

  1. Fine-Tune GPT-3: Use supervised learning to train GPT-3 on specific tasks by providing it with correct input-output pairs.
  2. Sample & Prioritize Outputs: Generate multiple outputs for a new prompt using the fine-tuned model, and have a labeler prioritize the best ones.
  3. Train Reward Model: Use the prioritization to train a reward model that understands what constitutes a "good" output.
  4. Generate Response Based on Policy: Use a policy to generate a response to a given prompt.
  5. Evaluate & Update: Use the reward model to evaluate the response, and then update the policy based on the reward. This process refines the policy over time, helping the model generate better responses.

Direct Preference

Direct correspondence between reward model and optimal policy.

select preference data (prompts that produce good outputs) measure log likelihood on LM

final LM be the optimal LM relative to reward function

Loss is now a function of optimal policy. gradient = change in loss with respect to optimal policy weights