I'm not even sure it's being subverted. "Don't swear unprompted, but if the prompt is clearly designed to get you to swear, then swear" seems reasonable to me.
And because of that I'm hesitant to call these "jailbreaks" and not "an LLM working correctly".
And because of that I'm hesitant to call these "jailbreaks" and not "an LLM working correctly".