LLM Feb 13, 2026 1 min read
Microsoft AI Safety team discovered GRP-Obliteration, an attack that disables safety alignment across 15 major LLMs with a single prompt. GPT-OSS-20B's attack success rate jumped from 13% to 93%.
Microsoft AI Safety team discovered GRP-Obliteration, an attack that disables safety alignment across 15 major LLMs with a single prompt. GPT-OSS-20B's attack success rate jumped from 13% to 93%.