RESEARCH

ZAYA1-8B Technical Report

ArXiv cs.AI · Fri, 08 May 2026 04:00:00 GMT

arXiv:2605.05365v1 Announce Type: new Abstract: We present ZAYA1-8B, a reasoning-focused mixture-of-experts (MoE) model with 700M active and 8B total parameters, built on Zyphra's MoE++ architecture. ZAYA1-8B's core pretraining, midtraining, and supervised fine-tuning (SFT) were

Read original source Discuss with A.S.I.S