QUOTE POST
#196Rishabh Agarwal@AGARWL_
And see the tweet thread from @KushaSareen here
Can LLMs adapt continually without losing base skills? Fast-Slow Training (FST) pairs "slow" weights with "fast" context. FST vs. RL: • 3x more sample-efficient • Higher performance ceiling • Less KL drift (better plasticity) • Continual learning: succeeds where RL stalls
3:37 PM · May 13, 2026 · 113K Views
12:24 AM · May 15, 2026 · 4.7K Views