LLM Reddit 3h ago 1 min read
A busy LocalLLaMA thread followed David Noel Ng’s RYS II results, which argue that repeated mid-stack transformer layers can still improve Qwen3.5-27B and that hidden states may align more by meaning than by surface language.