Skip to main content

Pretrain Alignment

Train LLM three steps

Pre-train -> Supervised Fine-tuning -> RLHF

  • Alignment = Supervised Fine-tuning + RLHF = Fine-tuning

Alignment

Models Naming Tips

  • Models with base mean only pre-train e.g. Llama-2-7b-base
  • Models with chat , instruct mean with alignment e.g. Llama-2-7b-chat

Alignment using Datasets

Knowledge Distillation

Alignment before and after

  • Paper:
  1. The Unlocking Spell on Base LLMs

Alignment different methods

  1. Response Tuning
  2. Rule-base adapter

Self-Alignment

Pretrain

Efficient Pretrain

DataSet

DataSet Quality 重要性

說明

Textbook 是 chatGPT 生成的 may be effected the result

Rephrasing the Web

Alignment 的極限