LLM X/Twitter Mar 19, 2026 2 min read

OpenAI said on March 6, 2026 that Codex Security is entering research preview for ChatGPT Pro, Enterprise, Business, and Edu users in Codex web. The company says the application-security agent uses project-specific threat models, contextual validation, and patch proposals, and in beta scanned more than 1.2 million commits.

LLM Reddit Mar 19, 2026 2 min read

A LocalLLaMA thread on March 18, 2026 pushed fresh attention toward Mamba-3, a new state space model release from researchers at Carnegie Mellon University, Princeton, Cartesia AI, and Together AI. The project shifts its design goal from training speed to inference efficiency and claims prefill+decode latency wins over Mamba-2, Gated DeltaNet, and Llama-3.2-1B at the 1.5B scale.

LLM Mar 18, 2026 2 min read

Google introduced Gemini 3.1 Flash-Lite on March 3, 2026 as its fastest and most cost-efficient Gemini 3 series model. The model is rolling out in preview through the Gemini API in Google AI Studio and Vertex AI, with pricing of $0.25/1M input tokens and $1.50/1M output tokens, plus claims of a 2.5x faster Time to First Answer Token and 45% higher output speed than 2.5 Flash.

LLM X/Twitter Mar 18, 2026 2 min read

OpenAI said on March 17, 2026 that GPT-5.4 mini is now available in ChatGPT, Codex, and the API, with a follow-up post confirming GPT-5.4 nano in the API. OpenAI's developer docs position mini as its strongest mini model yet for coding, computer use, and subagents, while nano is framed as the cheapest GPT-5.4-class model for high-volume tasks like ranking, extraction, and sub-agent work.