Models & Algorithms🇰🇷 한국어

SDFT: Learning Without Forgetting via Self-Distillation

No complex RL needed. Models teach themselves to learn new skills while preserving existing capabilities.

SDFT: Learning Without Forgetting via Self-Distillation

SDFT: Learning Without Forgetting via Self-Distillation

No complex RL needed. Models teach themselves to learn new skills while preserving existing capabilities.

TL;DR

  • Problem: Traditional SFT causes catastrophic forgetting when learning new tasks
  • Solution: SDFT (Self-Distillation Fine-Tuning)
🔒

Sign in to continue reading

Create a free account to access the full content.

Related Posts