* The LLM has strong and deep rooted belief in its knowledge (that a seahorse emoji exist).
* It attempts to express that concept using language (including emojis) but the language is so poor and inaccurate at expressing the concept that as it speaks it keeps attempting to repair.
* It is trained to speak until it has achieved some threshold at correctly expressing itself so it just keeps babbling until the max token threshold triggers.
(Edit: There is another long thread that contains an image that I thought was the seahorse emoji (although apparently the seahorse emoji doesn't exist...but i thought this was it so I don't know what is going on...) https://www.reddit.com/r/Retconned/comments/1di3a1m/comment/...)
And those text got into the training set: https://www.reddit.com/r/MandelaEffect/comments/qbvbrm/anyon...
llamasushi•52m ago
Explains why RL helps. Base models never see their own outputs so they can't learn "this concept exists but I can't actually say it."
bombcar•40m ago
bravura•21m ago
Example: "Is there a lime emoji?" Since it believes the answer is no, it doesn't attempt to generate it.