Overview:
- Problem: Models are big and it's hard to change facts that they learn.
- In a nutshell: mend generates a weight delta that's decomposed into a low-rank matrix like LoRA.
- "local, reliable, and general"
- "Local" means unrelated output is not changed. "Reliable" means the model takes the desired corrections. "General" meaning variations on similar questions which would need correction also are corrected.
- Works even on very large models.
Differences to Prior Art:
- ENN encodes editability into parameters of the model itself. MEND provides editability through independent model. (ENN is closer to fine-tune? MEND closer to LoRA?)
- KE uses raw edit example as input and produces a single rank-1 mask and rank-1 offset over fine tune grad. MEND maps model grads into model edits.