
Llama 4: Meta's first natively multimodal mixture-of-experts (MoE) architecture models
Meta has released the Llama 4 models—Scout and Maverick—featuring native multimodality, mixture-of-experts architecture, and state-of-the-art context lengths. Meta also previewed Behemoth, a trillion-parameter model that supposedly outperforms GPT-4.5 and Claude 3.7 Sonnet on STEM benchmarks.