On March 16, 2026, a r/LocalLLaMA link to Mistral Small 4 reached 504 points and 196 comments. The Hugging Face model card describes a 119B MoE with 4 active experts, 256k context, multimodal input, and per-request reasoning control.
#mistral
RSS FeedLLM Reddit Mar 17, 2026 2 min read
AI Mar 15, 2026 2 min read
Mistral has published Voxtral Realtime and Voxtral Mini Transcribe V2, adding sub-200ms streaming transcription, 13-language support, and open weights for the realtime model. The company also paired the launch with an audio playground in Mistral Studio and aggressive API pricing at $0.003/min and $0.006/min.
LLM Mar 8, 2026 1 min read
Mistral has launched Mistral 3, a new open multimodal family with dense 14B, 8B, and 3B models under Apache 2.0, plus a larger Mistral Large 3. The company says the lineup was trained from scratch and tuned for both Blackwell NVL72 systems and single-node 8xA100 or 8xH100 deployments.