5d ago

Marin releases Delphi scaling suite for pretraining predictions

0

Marin introduced its Delphi scaling suite to support reliable performance predictions during pretraining of open models. Researchers trained multiple small Dyna models from 72 million to 6.9 billion parameters under one fixed recipe and fitted a scaling law. The law extrapolated accurately to a 25-billion-parameter model on 600 billion tokens at roughly 1e23 FLOPs, matching observed Paloma macro loss within 0.2 percent error and confirming the fit across more than two orders of magnitude in compute.

Original post

To train better open models, we need predictable scaling. Delphi is Marin’s first step: we pretrained many small models with one recipe, then extrapolated 300× to predict a 25B-param / 600B-token run with just 0.2% error. Getting there took some work 🧵

12:25 PM · May 11, 2026 View on X
Reposted by

Really incredible work!

Will HeldWill Held@WilliamBarrHeld

To train better open models, we need predictable scaling. Delphi is Marin’s first step: we pretrained many small models with one recipe, then extrapolated 300× to predict a 25B-param / 600B-token run with just 0.2% error. Getting there took some work 🧵

7:25 PM · May 11, 2026 · 133.5K Views
7:27 PM · May 13, 2026 · 13.1K Views

Marin’s Delphi scaling suite is out!

With the right scaling recipe, small runs predicted a 1e23 FLOP run within 0.2%, extrapolating 300× past the largest run in the fit.

Will HeldWill Held@WilliamBarrHeld

To train better open models, we need predictable scaling. Delphi is Marin’s first step: we pretrained many small models with one recipe, then extrapolated 300× to predict a 25B-param / 600B-token run with just 0.2% error. Getting there took some work 🧵

7:25 PM · May 11, 2026 · 133.5K Views
8:03 PM · May 11, 2026 · 2.4K Views

Delphi changes how we evaluate new ideas: start small, sweep the param/token tradeoff, scale the key hypers, compare against forecasts, repeat.

David HallDavid Hall@dlwh

Marin’s Delphi scaling suite is out! With the right scaling recipe, small runs predicted a 1e23 FLOP run within 0.2%, extrapolating 300× past the largest run in the fit.

8:03 PM · May 11, 2026 · 2.4K Views
8:03 PM · May 11, 2026 · 177 Views

Also, Will is underselling the blog post. The interactive figures are excellent: they make the scaling intuition concrete, including what transfers and what breaks.

Worth reading https://openathena.ai/blog/delphi/

David HallDavid Hall@dlwh

Delphi changes how we evaluate new ideas: start small, sweep the param/token tradeoff, scale the key hypers, compare against forecasts, repeat.

8:03 PM · May 11, 2026 · 177 Views
8:03 PM · May 11, 2026 · 158 Views

scaling laws are beautiful and this is the best resource to understand them

openathena.ai
/blog/delphi/
Will HeldWill Held@WilliamBarrHeld

To train better open models, we need predictable scaling. Delphi is Marin’s first step: we pretrained many small models with one recipe, then extrapolated 300× to predict a 25B-param / 600B-token run with just 0.2% error. Getting there took some work 🧵

7:25 PM · May 11, 2026 · 133.5K Views
8:47 PM · May 11, 2026 · 70.5K Views

To train better open models, we need predictable scaling.

Delphi is Marin’s first step: we pretrained many small models with one recipe, then extrapolated 300× to predict a 25B-param / 600B-token run with just 0.2% error.

Getting there took some work 🧵

7:25 PM · May 11, 2026 · 133.5K Views
Marin releases Delphi scaling suite for pretraining predictions · Digg