3d ago

Pranav Shyam calls for moratorium on new optimizers

0

Pranav Shyam called for a moratorium on new optimizers until existing methods are better understood. The call follows Shampoo displacing Adam variants and the reparameterization of Shampoo with beta2 set to zero, now called Muon. Muon has spawned dozens of follow-on variants with only marginal gains. Rohan Anil described the pattern of incremental changes after Muon. Keller Jordan observed that a benchmark intended to curb such proliferation has not proven effective.

Original post

Moratorium on new optimizers until we figure out whats going on

4:19 PM · May 13, 2026 View on X

@zacharynado @sasuke___420 I remember this table! Though to be fair, Shampoo and KFac are in there too

Zachary NadoZachary Nado@zacharynado

the sloptimizer field is just getting started with shampoo and muon gen algorithms, the graveyard of adam variants got so bad you can't list them all on a page

12:05 AM · May 15, 2026 · 32.5K Views
6:08 AM · May 15, 2026 · 1.1K Views

We did Shampoo so we kill all the variants of Adam. Now we call Shampoo b2=0.0 as Muon, then created 10s of variants thats marginally better.

Pranav ShyamPranav Shyam@recurseparadox

Moratorium on new optimizers until we figure out whats going on

11:19 PM · May 13, 2026 · 19.2K Views
11:22 PM · May 13, 2026 · 17.4K Views

@recurseparadox This was what my benchmark was supposed to fix. But I’m not sure it’s working

Pranav ShyamPranav Shyam@recurseparadox

Moratorium on new optimizers until we figure out whats going on

11:19 PM · May 13, 2026 · 19.2K Views
4:35 AM · May 14, 2026 · 313 Views

the sloptimizer field is just getting started with shampoo and muon gen algorithms, the graveyard of adam variants got so bad you can't list them all on a page

Pranav ShyamPranav Shyam@recurseparadox

Moratorium on new optimizers until we figure out whats going on

11:19 PM · May 13, 2026 · 19.2K Views
12:05 AM · May 15, 2026 · 32.5K Views

from @frankstefansch1's https://arxiv.org/abs/2007.01547

Zachary NadoZachary Nado@zacharynado

the sloptimizer field is just getting started with shampoo and muon gen algorithms, the graveyard of adam variants got so bad you can't list them all on a page

12:05 AM · May 15, 2026 · 32.5K Views
12:05 AM · May 15, 2026 · 1.5K Views
Alexander DoriaAlexander Doria@Dorialexander

maybe we have been too fast to accuse auto-research. just primary male urge to optimize.

10:36 AM · May 15, 2026 · 10.3K Views
1:00 PM · May 15, 2026 · 11K Views

@zacharynado slopGD is crazy 😭😭

Zachary NadoZachary Nado@zacharynado
1:00 PM · May 15, 2026 · 11K Views
1:04 PM · May 15, 2026 · 596 Views

A big take away from @jeankaddour and Oscar’s “No Train No Gain” paper (https://arxiv.org/abs/2307.06440, NeurIPS 2023) is that Adam with LR decay is extremely difficult to beat, and many complex training strategies do not really buy you anything

Zachary NadoZachary Nado@zacharynado

the sloptimizer field is just getting started with shampoo and muon gen algorithms, the graveyard of adam variants got so bad you can't list them all on a page

12:05 AM · May 15, 2026 · 32.5K Views
8:49 PM · May 15, 2026 · 2.5K Views

maybe we have been too fast to accuse auto-research. just primary male urge to optimize.

Zachary NadoZachary Nado@zacharynado

the sloptimizer field is just getting started with shampoo and muon gen algorithms, the graveyard of adam variants got so bad you can't list them all on a page

12:05 AM · May 15, 2026 · 32.5K Views
10:36 AM · May 15, 2026 · 10.3K Views

5yo seeing a hole on the beach: what if we digged further. 20-30 something seeing a loss valley:

Alexander DoriaAlexander Doria@Dorialexander

maybe we have been too fast to accuse auto-research. just primary male urge to optimize.

10:36 AM · May 15, 2026 · 10.3K Views
1:07 PM · May 15, 2026 · 1.3K Views

@typedfemale Join my cause

Pranav ShyamPranav Shyam@recurseparadox

Moratorium on new optimizers until we figure out whats going on

11:19 PM · May 13, 2026 · 19.2K Views
6:35 PM · May 14, 2026 · 370 Views
Pranav Shyam calls for moratorium on new optimizers · Digg