r/LocalLLaMA reacted because this was not a polished game pitch. The hook was a local world model turning photos and sketches into a strange little play space on an iPad.
#world-models
RSS FeedThe paper drew attention because it challenges today’s data appetite, but the comments quickly tested the comparison to children.
HY-World 2.0 turns text, images, multi-view inputs, or video into 3D Gaussian Splatting scenes. The stronger signal is reproducibility: the authors say model weights, code, and technical details are available.
Google DeepMind introduced D4RT on January 22, 2026 as a unified model for dynamic 4D scene reconstruction and tracking. The company says it runs 18x to 300x faster than prior methods and is efficient enough for real-time applications in robotics and augmented reality.
Runway introduced Characters on March 9, 2026, a real-time video agent API built on GWM-1. The company says developers can create and control custom conversational avatars from a single image without fine-tuning.
Runway has raised $315 million in Series E financing led by General Atlantic. The company says the capital will fund next-generation world-model pretraining and expansion into new products and industries.
Google DeepMind posted on 2026-02-25 about Project Genie and linked a Q&A on world models. The post frames world models as environment simulators for agent training, education, and interactive media use cases.
Google announced Project Genie on 2026-01-29 and started rolling out access to Google AI Ultra subscribers in the U.S. (18+). The Google Labs prototype combines Genie 3, Nano Banana Pro, and Gemini for world sketching, exploration, and remixing workflows.
Google DeepMind announced Genie 3, a world model that generates interactive environments from text or image prompts. The system targets 720p at 24fps and sustains coherent interactive worlds for over one minute.