I love love love this piece – even just the tweet thread! Folks spend a lot of time talking about "alignment," the idea that we need AI values to agree with the values of humankind. The thing is, though, people have a lot of different opinions.
For example, if we made AI choose between democracy and the economy, it's 150% on the side of democracy. People are a little more split, and it changes rather drastically between different countries.
It's a really clear example of bias, but (importantly!) not in a way that's going to make anyone feel threatened by having it pointed out. Does each county have to build their own LLM to get the "correct" alignment? Every political party? Does my neighborhood get one?
While we all know in our hearts that there's no One Right Answer to values-based questions, this makes the issues a little more obvious (and potentially a little scarier, if we're relying on the LLM's black-box judgment).
You can visit Towards Measuring the Representation of Subjective Global Opinions in Language Models to see their global survey, and see how the language model's "thoughts and feelings" match up with those of the survey participants from around the world.
I wasn't there, but I cite this tweet like every hour of every day.
This is basically GPT's reaction to whatever simple, specific instructions you give it instructions.
The time they don't allow AI-generated images is when it would displace stock photography, explicitly saying this is because photographers make money that way.
Like six hundred of the rotating light emoji go right here. There's an assumption that going from unstructured->structured data is easy-peasy, no need to hallucinate anything, it's basically fancy regex... but that's not the case! LLMs are more than happy to make things up, even when they don't need to.