HN picked up Nanocode, an open JAX project that packages tokenizer training, pretraining, synthetic data generation, agentic SFT, and DPO into an end-to-end recipe for building a coding model on TPU infrastructure.
#open-source
RSS FeedA Show HN thread highlighted GuppyLM, a tiny 8.7M-parameter transformer with a 60K synthetic conversation dataset and Colab notebooks. The point is not state-of-the-art performance, but making the full LLM pipeline inspectable from data generation to inference.
A fresh LocalLLaMA thread argues that some early Gemma 4 failures are really inference-stack bugs rather than model quality problems. By linking active llama.cpp pull requests and user reports after updates, the post reframes launch benchmarks as a full-stack issue.
Netflix’s VOID reached Reddit as an open research release aimed at removing objects from video and repairing the interactions those objects caused in the scene. The notable details are the CogVideoX base, a two-pass pipeline, Gemini+SAM2 mask generation, and a 40GB+ VRAM requirement.
A Hacker News post highlighted OpenMed’s end-to-end protein AI pipeline, from ESMFold and ProteinMPNN to CodonRoBERTa, plus a 25-species model suite trained in 55 GPU-hours. The thread praised the engineering detail but also raised the usual questions about biological validation.
Lemonade packages local AI inference behind an OpenAI-compatible server that targets GPUs and NPUs, aiming to make open models easier to deploy on everyday PCs.
Cohere has entered the speech stack race with Transcribe, a 2B Apache 2.0 ASR model for 14 languages. Open weights, Hugging Face distribution, and a claimed 5.42 average WER headline the release.
A March 2026 Hacker News post reached 252 points and 261 comments around George London’s argument that coding agents could make free software relevant again. The core claim is that agents turn source-code access from a symbolic programmer right into a practical capability for ordinary users who need software changed on their behalf.
NVIDIA announced Dynamo 1.0 on March 16, 2026 as a production-grade open-source layer for generative and agentic inference. The release matters because it ties Blackwell performance gains, lower token economics and native integration with major open-source frameworks into one operating model.
A March 1 r/MachineLearning post compared 94 LLM endpoints across 25 providers and argued that open models were closing to within a single-digit quality gap of top proprietary systems. The real takeaway is operational: model choice is now about intelligence, price, speed, and deployment freedom at the same time.
Mistral announced Mistral Small 4 on March 16, 2026 as a single open model that combines reasoning, multimodal input, and agentic coding. Key specs include 119B total parameters, 6B active parameters per token, a 256k context window, Apache 2.0 licensing, and configurable reasoning effort.
Mistral introduced Leanstral on March 16, 2026 as an open-source code agent built specifically for Lean 4. The release combines 6B active parameters, an Apache 2.0 license, a new FLTEval benchmark, and immediate availability in Mistral Vibe, API form, and downloadable weights.