We're meant to assume correct sentences were written by humans and AI adds glaring factual errors. I don't think it is possible at this point to tell a single human written sentence from an AI written sentence with no other context and it's dangerous to pretend it is this easy.
Several of the AI images included obvious mistakes a human wouldn't have made, but some of them also just seemed like entirely plausible digital illustrations.
Oversimplifying generative AI identification risks overconfidence that makes you even easier to fool.
Loosely related anecdote: A few months ago I showed an illustration of an extinct (bizarre looking) fish to a group of children (ages 10-13ish). They immediately started yelling that it was AI. I'm glad they are learning that images can be fake, but I actually had to explain that "Yes, I know this is not a photo. This animal is long extinct and this is what we think it looked like so a person drew it. No one is trying to fool you."
(minor spoiler)
The text accompanying an image of a painting:
> This image shows authentic human photography with natural imperfections, consistent lighting, and realistic proportions that indicate genuine capture rather than artificial generation. Meindert Hobbema. The Avenue at Middelharnis (1689, National Gallery, London)
I don't mind that you're selling an AI product if it's good but at least put some humanity on the marketing side.
I don't think this is accurate. AI has a flavour or tone we all know, but it could have generated factually plausible statements (that you could not diagnose in this test) or plausible text.
I could not tell the real from fake music at all.
I support (and pay for) Kagi, but wasn't overly impressed here. At worst I think it might give people too much confidence. Wikipedia has a great guideline on spotting AI text and I think the game here should integrate and reflect its contents: https://en.wikipedia.org/wiki/Wikipedia:Signs_of_AI_writing
- AI slop is trivially factually wrong, and frequently overconfident.
- AI slop is verbose.
But, as you note, IRL this is not usually the case. It might have been true in the GPT-3.5 or early GPT-4 days, but things have moved on. GPT-5.1 Pro can be laconic and is rarely factually wrong.
The best way to identify AI slop text is by their use of special and nonstandard characters. A human would usually write "Gd2O3" for gadolinium oxide, whereas an AI would default to "Gd₂O₃". Chat-GPT also loves to use the non-breaking hyphen (U+2011), whereas all humans typically use the standard hyphen-minus character (U+002D). There's more along these lines. The issue is that the bots are too scrupulously correct in the characters they use.
As for music, it can be very tough to distinguish. Interestingly, there are some genres of music that are entirely beyond the ability of AI to replicate.
Sounds interesting, what are some of those genres?
I started on "Level 1" and got 2 things wrong (both false positives if it matters) and instead of feeling like I learned anything, I felt as though I was set up to fail because the image prompt was missing sufficient context or the text prompt was too simple to be human. Either I was dumb or the game was dumb.
Maybe I'm just too old and 8-11 year-old kids wouldn't be so easily discouraged, but I'd recommend:
1. Picking on one member of the "slop syndicate" at a time.
2. Show some examples (evidence) before beginning the evaluation.
>This was actually AI-generated slop! Repeats 'water is wet' multiple times.
I didn't know writing "water is wet" repeatedly was enough to de-humanize you.
>In many situations, it could be argued that grass may sometimes appear to have a greenish quality, though this might not always be the case.
>This was actually AI-generated slop! Won't commit to 'grass is green' and uses uncertain words.
What? Not all grass is green.
Fun times ahead.
Der_Einzige•39m ago
https://arxiv.org/abs/2510.15061
Also somewhat tangentially relevant video: https://www.youtube.com/watch?v=Tsp2bC0Db8o