Machine Learning Research
Art Attack: ArtPrompt, a technique that exploits ASCII art to bypass LLM safety measures
Seemingly an innocuous form of expression, ASCII art opens a new vector for jailbreak attacks on large language models (LLMs), enabling them to generate outputs that their developers tuned them to avoid producing.