The paper is here. It's sadly not about AI detection, but rather whether large language models have a model of the world or are just faking it. If you come in thinking it's the former you're rather quickly brought to your senses:
Do large language models (LLMs) have beliefs? And, if they do, how might we measure them?