AI Red Teams Are Standardizing on Structured Output Attacks
Attackers are no longer just trying to jailbreak a model’s text—they’re targeting the JSON, XML, and function-call formats that modern AI systems trust downstream. Security teams need to understand how structured outputs can silently turn a harmless-looking response into unsafe automation or data leakage.