I honestly though that ASCII art didn't work that well for LLMs! But maybe they're just bad at generating it, not reading it? In this case, the semantics of building a bomb makes it through the alignment force field:
And yeah, it's still bad at generating ASCII art. So at least we can still employ humans for one thing.
I posted this instead of the article itself because maybe a discussion is better than a press release.
Amazon started using AI for support a while back. The only time I've had to interact with it: it asked the right questions, came to the appropriate course of action, told me the correct things, and then did the exact opposite on the backend which then required me to track down an actual human (quickly successfully I might add) to fix it.
And everyone's dream:
I pray for the day I can get an LLM representing a bank with apparent authority to negotiate financial arrangements.
Governments buy into hype same as everyone else.
Even if it works in the short-term news outfits are going to get screwed. There's no long-term money for Google in journalism.