Zyphra released the first MoE diffusion model with a 7.7x speedup
Zyphra introduced ZAYA1-8B-Diffusion-Preview, the first-ever MoE diffusion model converted from a standard language model. The main result: inference became 7.7

Zyphra released ZAYA1-8B-Diffusion-Preview — the first MoE diffusion model successfully converted from an autoregressive language model. The model demonstrates that such transformation is possible without quality loss, while inference speeds up by 7.7x — a significant leap for performance.
What Happened
Typically, language models operate in autoregression mode: generating words sequentially, one after another. This is slow because each step depends on the previous one — processing cannot be parallelized. Zyphra redesigned ZAYA — an MoE model (Mixture of Experts), which selects different neural subnets for different inputs — into a discrete diffusion model. In diffusion, the generation logic is entirely different: the model takes a noisy representation and iteratively cleans it, processing multiple computation layers in parallel. The idea is not new — diffusion works well for images and text. But redesigning an MoE architecture from an autoregressive paradigm to a diffusion one while preserving quality — this is something previous attempts had not achieved so cleanly.
Why This Speeds Things Up
The key lies in which GPU resource each mode uses. Two concepts are important here:
- Memory-bandwidth bound tasks: read lots of data from memory, process little. Autoregressive generation — a classic example: you keep the entire context, reload it with each token, add a new token, generate the next one
- Compute-bound tasks: read data once, process many times. Diffusion — multiple iterations of cleaning the same tensor, each iteration requires a full pass through the neural network
- GPU architecture: modern GPUs grow faster in FLOPS (compute power) than in memory bandwidth. You have many compute cores, but they often wait for memory to deliver data
Transitioning ZAYA from memory-bound to compute-bound means GPU compute cores work closer to maximum load. Hence the 7.7x speedup.
Metrics Remained Unchanged
Zyphra tested the quality of the diffusion version against the original autoregressive ZAYA1-8B. Scores remained at the same level — the model did not lose its ability to generate text, recognize context, or follow instructions. This is not obvious: often when transitioning between paradigms, something degrades. Not here. The result means that the diffusion approach and MoE architecture are compatible, and reconversion does not destroy the knowledge the model accumulated during training.
What It Means
MoE-diffusion models are moving from laboratories into practical tools. For companies, this means: you can take an existing MoE model and get 7-8x inference speedup without retraining and without new GPUs.