Zyphra Releases ZAYA1-8B: A Reasoning MoE Trained on AMD Hardware That Punches Far Above Its Weight Class
MarkTechPost1 min read
Read Full Article at MarkTechPost →
Ad Slot — In-Article (728x90)
Zyphra releases ZAYA1-8B, a reasoning Mixture of Experts model with only 760M active parameters that outperforms open-weight models many times its size on math and coding benchmarks — closing in on DeepSeek-V3. 2 and surpassing Claude 4.
5 Sonnet on HMMT'25 with its novel Markovian RSA test-time compute method. Trained end-to-end on AMD Instinct MI300 hardware and released under Apache 2. 0, it sets a new standard for intelligence density in the small language model weight class.
This is a summary. For the full story, read the original article at MarkTechPost.
Original source: MarkTechPost