Meta introduced two new multimodal open-weight models: Llama 4 Scout and Llama 4 Maverick, built on a mixture-of-experts (MoE) architecture.
Llama 4 Scout features 17 billion active parameters and 16 experts, optimized for single-GPU deployment, making it suitable for edge and enterprise applications.
Llama 4 Maverick, with 128 experts, is designed for heavy-duty workloads and large-scale use cases. Meta claims Maverick outperforms OpenAI’s GPT-4o and Google’s Gemini 2.0 on coding, reasoning, multilingual, long-context, and image benchmarks, and is competitive with the much larger DeepSeek V3.1.
MoE architecture enables efficiency by activating only relevant parts of the model for each input, reducing computational overhead.