1d ago

AsymFlow achieves 1.57 FID on ImageNet in pixel space

0

Hansheng Chen introduces AsymFlow, a flow-based image generation method that retains velocity information in a low-rank subspace instead of using JiT x0-prediction. The model operates directly in pixel space without a VAE and records an FID of 1.57 on ImageNet, the highest among published pixel-space flow models. When used to finetune FLUX.2 klein, the resulting checkpoint outperforms the base model on HPSv3, DPG, and GenEval, ranking first overall on HPSv3 while delivering sharper textures and roughly 40 percent faster inference. The work was shared by researchers including Kosta Derpanis at York University.

Original post

FLUX.2 Klein in pure pixel space! No VAE. AsymFlow - hyper-realistic images by working directly in pixel space rather than using compressed latent representations. - sharper textures, superior visual fidelity - 40% faster - low-rank noise parameterization to solve high-dimensional bottlenecks ComfyUI support incoming https://hanshengchen.com/asymflow/

5:08 AM · May 14, 2026 View on X
Reposted by

Cool work by my former lab at Stanford on pixel-space image diffusion!

Hansheng ChenHansheng Chen@HanshengCh

New paper: AsymFlow🔥 JiT x0-prediction is not enough for pixel generation. Better keep velocity in a low-rank subspace: - 1.57 FID on ImageNet (best pixel flow model) - Finetunes FLUX.2 klein into pixel space, beats the original on HPSv3/DPG/GenEval (#1 overall on HPSv3) 1/7

1:58 PM · May 14, 2026 · 43.8K Views
4:55 PM · May 14, 2026 · 8.5K Views
AsymFlow achieves 1.57 FID on ImageNet in pixel space · Digg