Cohere Launches Tiny Aya: 3.35B Open-Weight Models Supporting 70+ Languages for Offline Use
Overview
Cohere unveiled Tiny Aya at the India AI Summit on February 17, 2026 — a family of compact, open-weight multilingual models designed to run offline on standard laptops. The release targets language accessibility in regions underserved by English-centric AI tools.
Model Specifications
- Parameters: 3.35 billion
- License: Open-weight (MIT)
- Training infrastructure: Single cluster of 64 H100 GPUs
- Languages supported: 70+
Regional Variants
Cohere released region-specific fine-tuned variants:
- TinyAya-Fire: South Asian languages — Hindi, Urdu, Bengali, Punjabi, Gujarati, Tamil, Telugu, Marathi
- TinyAya-Earth: African languages
- TinyAya-Water: Asia-Pacific, Western Asian, and European languages
Availability
Models are available on HuggingFace, Kaggle, and Ollama for local deployment, as well as through the Cohere platform API. The efficient training footprint — just 64 H100 GPUs — also positions Tiny Aya as a reference point for cost-effective multilingual model development.
Source: TechCrunch
Related Articles
Why it matters: inference cost is now a product constraint, not only an infrastructure problem. Cohere said its W4A8 path in vLLM is up to 58% faster on TTFT and 45% faster on TPOT versus W4A16 on Hopper.
Why it matters: document agents fail when PDF parsing destroys table and column structure. LiteParse uses a monospace grid projection approach instead of heavy layout models, and the code is open source.
r/MachineLearning did not reward this post for frontier performance. It took off because a 7.5M-parameter diffusion LM trained on tiny Shakespeare on an M2 Air made a usually intimidating idea feel buildable.
Comments (0)
No comments yet. Be the first to comment!