ZAYA achieves quality at parity with Qwen3-4B and Gemma3-12B, while exceeding models such as IBM-Granite-4-H-Tiny.
ZAYA’s MoE architecture enables the response times of an 800M dense model, but the quality of a 12B dense model. Quality is never sacrificed for performance.
First AI model trained entirely end-to-end on AMD’s hardware, software, and networking stack.