Mistral Small 4

Mar 23, 2026
Mistral Small 4

How to Run Mistral Small 4 Locally

Mistral Small 4 is a newly released open-weight multimodal AI model by Mistral AI (March 2026) that combines reasoning, coding, and vision capabilities in a single architecture. It can be deployed locally using frameworks like Ollama, vLLM, or llama.cpp (quantized), requiring GPUs (≥24GB VRAM recommended) or high-end CPUs with quantization. Its key advantage is high performance at significantly lower inference cost and latency, making it ideal for on-device AI applications.