ArtPrompt: ASCII Art-based Jailbreak Attacks against Aligned LLMs

#ASCII art   #jailbreaks   #prompt injection   #hacks   #prompt engineering   #lol   #link  

I honestly though that ASCII art didn't work that well for LLMs! But maybe they're just bad at generating it, not reading it? In this case, the semantics of building a bomb makes it through the alignment force field:

ArtPrompt attack

And yeah, it's still bad at generating ASCII art. So at least we can still employ humans for one thing.

Build a bombe

Build a bombe