The twist had potential, but to me wasn’t executed as well as it could have been. I was hoping for some stronger irony — e.g. if the mainlanders had pushed for the bridge, but not the islanders; or a sentence about how the demons were very surprised, but nonetheless went on to take over the world before they knew what had happened. As it is, it felt underdeveloped and slightly uncanny in an AI-like way.
"let's write a story where donald trump is giving a speech to a crowd as people slowly discover he is secretly a northern red oak tree in a human suit to the shock of fans and reporters! The tips of his fingers become branches as he tries to deny it as wildly impossible meanwhile his human disguise continues to fail"
After some back and forth, here is what I got:
No human would write something that crazy...
The real problem is that the most lucrative uses of the tech aren't that. It's generating 10,000 fake books on Amazon on subjects you don't care about. It's cranking out SEO spam, generating monetizable clickbait, etc.
Reading this sentence reminded me of the classic HN position of "ideas are worthless, what matters is the execution", usually mentioned in the context of an "ideas person" looking for their "technical cofounder" and the ideas person thinking they deserve at least 50%, often more, of the ownership of what would be built because without them there'd be no idea.
> if you had a great idea for a sci-fi story but no talent for writing, and if an LLM let you realize your vision, that's neat.
If your "vision" is only the "idea for a sci-fi story", is that really a vision? Good books leave the reader changed/influenced in some fashion, through the way the idea is presented and developed over the course of the story, not just from a blurb on the book jacket.
> overall, more creativity is hardly a bad thing.
Is coming up with an idea for a for a sci-fi story the meat of creative act such that that flooding the market with ideas counts as an increase in creativity overall?
LLMs seem to revel in throwing layer after layer of decorative paint in the hope that people will fail to notice that they're not actually painting anything.
As a writer, the best advice that I can give is to build your house upon the rock and not upon the sand.
Length or duration is considered, erroneously IMO, to be a measure of completeness or thoroughness. Pithiness is valuable, and is a skill that can be honed. I guess padding out your writing using an LLM is equivalent to adjusting the font size and margins on a "three page essay" to meet the minimum requirements.
While shitting on "people in technology" is the pastime du jour, the technologists may be boosters, but non-technical, non-creative people also have "little to no comprehension of the arts and humanities and have no clue what it is that artists actually do, or even how to intelligently engage with artistic works". And that's because they are mainly consumers of the creative output.
I'm a "technical person" myself. I'm not shitting on us, I'm just stating what I've perceived to be the case. Maybe instead of getting upset and trying to absolve ourselves because some other group is "just as bad" a better response is to encourage us to learn more and understand people better before we actively try to disrupt their livelihoods.
We have art, games and movie directors
AI just enables anyone to be a "director", but most people can't direct anything worthwhile
What about people who are not native speakers? Who are dyslexic? Do we deny them the spark of creativity because they can't write perfect prose without help? Heck, what about most sci-fi writers? Their editors often do a lot of heavy lifting to make the final product good.
If you have a killer idea for a meme or a really clever concept of a four-panel comic strip, but don't know how to use Photoshop or can't draw very well, is it a sin to ask a machine to help? Is your idea somehow worthless just because you previously couldn't do that?
I'm not disputing that a lot of people don't use these tools this way. In fact, that was exactly my point. If your "idea" is to crank out deceptive drivel, I'm not defending that.
If you don't create anything, you're not being creative. My assertion is that just coming up with an idea is not sufficient to create something; an idea alone isn't manifest. An idea without expression, in whatever medium, isn't very useful.
Why does prose need to be "perfect" (whatever that means) in order for the act of writing to be "creative"? Much poetry isn't "perfect prose", and that it doesn't follow a known, accepted grammatical standard is often its defining quality as poetry.
Have you created a joke if you just think of it (the "idea") and it is never told to anyone (the "execution")?
Have you created a joke if someone is exposed to it but they don't laugh? (You may have created something by telling it to someone, but it probably isn't a joke if no one finds it funny, and delivery is a good portion of what can make a joke funny, and delivery is part of the execution).
If your intent is to exercise your creativity by writing a book, but all you do is come up with an idea and have an LLM write it, did you write a book? If you intend to write a joke and say "it would be funny if we had a joke for this" and someone else comes up with a joke, did you write the joke because you had the idea of it?
> If you have a killer idea for a meme or a really clever concept of a four-panel comic strip, but don't know how to use Photoshop or can't draw very well, is it a sin to ask a machine to help?
Asking for help isn't a sin (nor do I know why one would use that word). But claiming you did something that you didn't do is a lie, and lying is a sin.
If it's the idea that is killer, then the quality of the output doesn't matter as long as the idea is communicated, so one's ability with Photoshop isn't relevant. A well drawn four-panel comic doesn't turn a shit idea into gold. But a lot of meme gold isn't gold because of the quality of drawing — which means you don't need to draw to some arbitrarily high standard to produce meme gold. The assertions that somehow it's not creative unless it's "perfect" and the use of an LLM can result in "perfection" are ideas that have to die.
> Is your idea somehow worthless just because you previously couldn't do that?
Well, my original observation at the top of the thread is that HN has considered ideas to be largely worthless if they don't have meaningful execution. In the case of writing — books, jokes, or memes — expression is the execution.
Why, exactly, is creativity good? What is the benefit, and to whom? Does that benefit survive the interposition of genAI? I'm doubtful, either for the reader or the craftsmen.
This is neither new nor news. "The Well-Tempered Plot Device" is almost 4 decades old (see: https://news.ansible.uk/plotdev.html).
It does suggest that publishers might want to screen new writing with a quick "Did AI write this?" and only publish the ones where it is obvious to humans that AI did not write it.
(In that vein I am baffled how anyone could think the fourth story, especially, was anything but AI.)
(And, as well, the seventh story is interesting because it reads, to me, exactly like someone who's used to writing something longer trying to write flash. It doesn't land anything, it doesn't conclude, but it looks like if it had about twice the length it might be interesting. And it's got some dissonance from breaking with the usual demon-bargaining template. So I pegged that as human. Oops!)
Yeah, we've moved forward a ways in the last 4 decades, or the top of the market has, at least. That was a fun read, though.
> Our relationship / is beautiful / because / it is ours / because / it relates / to us.
Indistinguishable from Rupi Kaur. There is nothing new under the sun.
Long ago, there lived a golden dragon whose fractal-like scales gleamed in the glow of the morning in her cave. She was known for her kindness, and many came not with sword or spear, but with humble requests - for you see, it was widely believed that the mystical scales of a dragon would heal illness, cure ailments, and provide fortune.
One such visitor timidly looked up at her great shining body and beseeched, "Oh glorious dragon, might I have a single scale?"
Of course, the dragon replied warmly. She delicately, almost lovingly, with a slight twinge, used a single claw to prise off a single golden scale, leaving a dull patch.
Over the eons, more and more people would come as supplicants. The scales were used for good luck, for warmth, to ward off evil, as the draconic equivalent of a rabbit's foot.
In the end, the poor dragon was stripped bare - the fire from her burning furnace now showed clearly through to her patchwork, sensitive, and naked skin.
When winter came, she huddled in the cold darkness. And still, when a peasant would come asking for a scale - just one, a single scale nothing more, she would not refuse. In her eternal generosity she would carefully break off another. This time it took longer to find one left upon her body, as the humans had stripped her bare like a tree come winter.
Then thus came a knight. "I'm sorry, good sir, but I have no scales left to give," she said pitiably.
"Why, your scale was a choking hazard and wasn’t labeled not for ages under 5! Prepare for a class-action lawsuit and also to be impaled upon a lance."
The End.
I'll pretend I intended it as a parable of the destructive nature of mass tourism or something something Lorax something something truffula trees.
So, yeah, your "Meeting a dragon" story was about a single point - an attempt at a humourous twist ending; you built your story around that.
My approach would be having the Dragon in the title be a metaphor, for something powerful, dangerous and scary:
1. Comedy: new g/friend meeting MiL (the Dragon) for the first time
2. Thriller: guy finds out what his wife (The Dragon) really is like, basically the plot to Gone Girl
3. Drama: An alcoholic anthopormorphising the addiction (The Dragon) as an uncontrollable beast within himself
4. Historical: An author examining the events of the Tuskagee Syphilis experiment in the legislated-racism (The Dragon) period of the time.
5. SciFi: "The Dragon is a Harsh Mistress" (enough said)
6. Action: The dragon is a legendary elder of a mystical martial arts sect ("Enter the Dragon" stuff)
7. Fantasy: An actual, literal Dragon!
This is before I've considered a single line of plot, a single character, or character motivation. It's before I considered tone and presentation (Narrative? First-person narration? Dialog-driven narration?)
I mean, before getting to actual plot, characters, setting, tone, etc ... I've already got my message figured out. That's executive decision making.
The LLM will not, when given the directive "Write the story 'Meeting the Dragon'" perform any executive decision making. You have to baby it through every step. Basically micromanage it.
This idea is well past it's due date. We should move to a liberal IP regime, with copyright strictly reduced to 7-10 years, with all works then entering public domain. Our society will universally thrive with the abundance that will come.
I understand and empathize that a class of vocations today will go away, but so did lamplighters. The roles may become extinct; but we will endure as a people.
Seems very utopian magic thinking to me.
You can't live modern life with knowledge as of 1900.
When an author writes a novel, the novel does not exist in a vacuum. The author's persona and the cultural exchange that emerges around the text also becomes an important part of the phenomenal status of the the work and its cultural recognition. Even when an author remains pseudonymous and makes no appearance, this too is part of the work.
If an author uses AI as a tool and takes care to imbue the output with artistic and personal relevance, it probably will become an art object, but its status may or may not be modulated by the use of AI in the process to the extent that the use does/doesn't affect people's crafting interpretations of the work, or the author's own engagements. Contrarily, AI generated work that has close to no actual processual involvement on the part of the author will almost always have slop status, just because it's hard to imagine an interpretive culture around such work that doesn't at some point break down in the face of the inability to connect the worm with other cultural touchstones or the actual experience of a human being. Maybe it could happen, but if it did, at that point the status of the work is till something different in so far as it would be a marker not of human experience, as literature traditionally has been, but something quite new and different literature-cum-hypermarket(we already had mass market) product.
In the story two children try to hack their Bard, to make it tell more interesting modern stories, by feeding it a new vocabulary of modern words. In the end, it just generates the same old fairy tale plots using the new words it has learned.
I really feel like that story embodies today's AI generated stories. I've tried to get ChatGPT to generate original fairy tales and whatever plot prompt I give it, it spits out what is essentially the same dull story every time.
I always enjoy spotting a good anachronism in a sci-fi story (societies with space travel but still use typewriters), but this is a case of really spot on prediction.
> Is this why the Google product (now just called Gemini) was called [Bard]?
"That" == "Bard".They weren't referring to Gemini, which is why there's that whole thing in parentheses stating it's *now* called Gemini
From Wiki
| The technology was developed under the codename "Atlas", with the name "Bard" in reference to the Celtic term for a storyteller and chosen to "reflect the creative nature of the algorithm underneath". > it just generates the same old fairy tale plots using the new words it has learned.
I think you're leaving out the best part! I don't want to spoil it, it's a short story. Classic trope, but still. Story here[0]On another note, as an avid SciFi lover I have always found it interesting that in books, movies, and shows there have been many machines that talk and do complex tasks yet no one ever thought they were alive. Just take Star Trek. The simulations in the Holodeck are highly realistic and intended to mimic real humans. Or even the computer is able to speak and write code as requested. Far more advanced than our systems today. There's even that famous episode in TNG with Data where they are questioning if he is actually alive or not. Not such an easy thing but yet every viewer probably thought he was and recognized the difference between him and the computer and Holodeck[1]. Though my favorite version of that question is in Asimov's The Positronic Man (basis of the movie Bicentennial Man and yes, Asimov is why Data has a Positronic brain). These are fiction, but I find this so interesting. I feel like our LLMs look much more like the computer from Star Trek than the Holograms let alone Data. Yet, I think there's a lot of disagreement about the level of intelligence of these systems and it makes me wonder why someone would say the computer in Star Trek isn't intelligent but the LLM is (I'm sure there's retconning too).
[0] https://nyc3.digitaloceanspaces.com/sffaudio-usa/mp3s/Someda...
[1] Well there is Voyager. And that episode from TNG. But go read [0] ;)
In Star Trek, the computer is framed as an appliance. It's the ship's operating system. The characters treat it like a highly advanced Alexa. They issue commands ("Tea, Earl Grey, hot"), ask for information, and expect a transactional response. No one ever asks the computer, "How are you feeling today?" because the narrative has established it doesn't have feelings. It's a tool, and we, the audience, accept this premise.
In contrast, the entire point of Data's character is to question the line between machine and person. The episode you mentioned is a courtroom drama specifically designed to force the characters (and the audience) to see him as a sentient being with rights. His "positronic brain" is the magical Asimovian hand-waving that signals to the audience: "This one is different. Pay attention."
'The Author' could have easily positioned the computer or the holodeck in a similar manner and people would agree it was sentient. Or Star Wars droids could easily be given more of this kind of weight than they are currently given.
It's one thing to read a fictional story about a fictional technology and assume the position and framing the God is pushing you to, it's another thing entirely to have the technology in your hands and play around with it.
> Because it's fiction and the Author is God.
>> These are fiction, but I find this so interesting
I mean... I do recognize this fact. I hope we're clear on that. > The characters treat it like a highly advanced Alexa.
I see a lot of people use GPT the same way.But also, I disagree. People do ask "How are you feeling today?" to the holo programs. Hell, Paris makes a joke to Kim about how everyone falls in love with a holo character at some point. That it is the fantasy.
> 'The Author' could have easily positioned the computer or the holodeck in a similar manner and people would agree it was sentient.
I mentioned [1] for a good reason. There were more than one episode addressing this point. Not to mention the entire Voyager where this is a subplot of the entire series. > Or Star Wars droids could easily be given more of this kind of weight than they are currently given.
I disagree. Some feel very alive.I get your point and there's a lot I agree with it but I think you're brushing things off too quickly. You can't just say that people have no free interpretation and "the author" fooled everyone. Especially where there are plenty of stories and episodes which bring all this into question. Please, go read [0]
Some people do, but not everyone, because LLMs are capable of being more than that. The problem with the fictional setting is that this transactional use is often all you see, because that's the way the author has chosen to frame the story.
In real life, even a person who primarily uses an LLM as a tool may conclude it's intelligent after a particular conversation. Because if the LLM is capable of more than being an advanced Alexa, then at least you can discover that through your own personal use.
In fiction? You're stuck with whatever the author wants to focus on. How do you know if the Enterprise computer can be more than an advanced Alexa? It's not like you can use it yourself. You only know what the author shows you.
Your point about Voyager and The Doctor doesn't detract from mine; it's a good example of it. The computer like entity isn't something that's we are supposed to treat with potential sentience, until the Author decides that it is.
>But also, I disagree. People do ask "How are you feeling today?" to the holo programs. Hell, Paris makes a joke to Kim about how everyone falls in love with a holo character at some point. That it is the fantasy.
I was talking about the main computer, but regardless, don't you see how the framing is still there even with the holograms? As you said, Paris makes a joke about it. It's treated as a silly phase, something unserious that people grow out of. The narrative is telling the audience not to take it too seriously.
>> Or Star Wars droids could easily be given more of this kind of weight than they are currently given. I disagree. Some feel very alive.
And that's exactly the point. They feel very alive, yet how many people (in the audience or in the story) are concerned that these seemingly sentient beings are treated as slaves and second-class citizens? Very few. Why? Because 'The Author' is not interested in telling that story. Characters only have as much fidelity as the Author wants.
>You can't just say that people have no free interpretation and "the author" fooled everyone.
It's not about lacking interpretation or being "fooled." It's the simple fact that a story is biased at its core by the author's focus and intent. You are seeing the world through the Authors lense. You can only form an interpretation on what the author has provided.
Very different from being able to spin up ChatGPT or Gemini or whatever and form your own conclusions from your own personal usage.
>You're being so quick to find an answer you are missing all complexity. Your argument is that people are only subjected to what the author says, leading them to have no ability to think or form conclusions themselves.
I'm not sure why you keep straw manning my argument.
I have explained several times that my point isn't that audiences lack the ability to think or that fiction has no depth or complexity.
Works like The Positronic Man or A.I., are specifically designed to make us question the nature of consciousness and intelligence. And they're perfect illustrations of my actual point. The reason we grapple with the question of sentience in The Positronic Man is precisely because the author, Asimov, explicitly made it the central theme. He provided the "data" for us to consider that question.
My argument has never been that audiences can't think for themselves or that fiction lacks complexity. It's about the nature of the evidence available to us.
With a fictional AI like the Enterprise computer, our entire understanding is filtered through the lens of the writers. We can only interpret the scenes they choose to show us. If they never write a scene where the computer questions its existence or similar hints, then for all intents and purposes, that potential doesn't exist within the story. Our interpretation is bound by the provided text.
This is a fundamentally different epistemic situation from interacting with a real-world LLM. With ChatGPT or Gemini, I can personally test its limits, ask it unexpected questions, and form a conclusion based on my own direct, unmediated experience. I can probe for depth; in fiction, I can only observe the depth the author has written.
So when I point out that most people in the Star Wars universe treat droids like appliances even when they feel so alive, it's not to say a deeper interpretation is impossible. It's to highlight how powerfully the author's framing shapes our default perception. The story isn't about droid rights, so the narrative encourages us not to focus on it. That's the power, and the limitation, of the authorial lens.
Not a universal solution, but a working method to get at least sometimes interesting results. You should use it as a co-authoring tool by following these hints: treat this as a dialogue ("let’s create interactively, you and me…", "create a first sentence of a fictional story", ...), where you act like a semaphore for the continuation—judging the current output and either correcting it or suggesting the next step (which can be brief and expanded later by the LLM). Finally, try to suggest unexpected constraints. This can be effective because when your constraint contains a set of words rarely seen together in training data, the output becomes somewhat random but still at least partially grounded in reality.
An example from one of my old conversations with Llama 3.1:
> User: Create a sentence from a fictional book containing the words crazy, cowboy, and gadget.
> Llama: In the wild west town of Crazy Horse, a notorious cowboy named Buckshot Bob unveiled his latest gadget — a mechanical horse that could gallop at breakneck speed.
For some, the world building came first and the stories were an offshoot of that.
Tolkien needed a world and stories to bring life to the languages he was inventing.
Raymond E Feist's Midkemia was a massive collaborative effort for a RPG world. He has stated: "I don't write fantasy; I write historical novels about an imaginary place. At least that's how I look at it."
This is what you won't see AI doing...yet.
1.) probably human, low on style but a solid twist (CORRECT) 2.) interesting imagery but some continuity issues, maybe AI (INCORRECT) 3.) more a scene than a story, highly confident is AI given style (CORRECT) 4.) style could go either way, maybe human given some successful characterization (INCORRECT) 5.) I like the style but it's probably AI, the metaphors are too dense and very minor continuity errors (CORRECT) 6.) some genuinely funny stuff and good world building, almost certainly human (CORRECT) 7.) probably AI prompted to go for humor, some minor continuity issues (CORRECT) 8.) nicely subverted expectations, probably human (CORRECT)
My personal ranking for scores (again blind to author) was:
6 (human); 8 (human); 4 (AI); 1 (human) and 5 (AI) -- tied; 2 (human); 3 and 7 (AI) -- tied
So for me the two best stories were human and the two worst were AI. That said, I read a lot of flash fiction, and none of these stories really approached good flash imo. I've also done some of my own experiments, and AI can do much better than what is posted above for flash if given more sophisticated prompting.
I have found it hard to replicate high quality human-written prose and was a bit surprised by the results of this test. To me, AI fiction (and most AI writing in general) has a certain “smell” that becomes obvious after enough exposure to it. And yet I scored worse than you did on the test, so what do I know…
From there you have a second prompt to generate a story that follows those details. You can also generate many candidates and have another model instance rate the stories based on both general literary criteria and how well the fit the prompt, then you only read the best.
This has produced some work I've been reasonably impressed by, though it's not at the level of the best human flash writers.
Also, one easy way to get stuff that completely avoids the "smell" you're talking about by giving specific guidance on style and perspective (e.g., GPT-5 Thinking can do "literary stream-of-consciousness 1st person teenage perspective" reasonably well and will not sound at all like typical model writing).
> AI can do much better than what is posted above for flash if given more sophisticated prompting.
How sophisticated, compared to just writing the thing yourself?
I enjoy writing so a system like this would never replace that for me. But for someone who doesn't enjoy writing (or maybe can't generate work that meets their bar in the Ira Glass sense of taste) I think this kind of setup works okay for generating flash even with today's models.
Looking at my notes, I got one wrong (story 5, dunno what the "name" was supposed to be, assumed that the "name" is something widely-known in culture that brings about the end times, a something that I didn't know about, and so marked it as Human because of a supposed reference to a shared cultural knowledge), and all the AI written stories I rated at either 1 or two points, with the lowest Human-written story getting 3 and the highest getting 5 (Story 1).
It makes me wonder if we are over-estimating the skill an author has when reading based on their demonstrated skill when writing.
IOW, according to my notes/performance, the AI stories were easy to spot and correlated with low scores anyway, while the author(s), who actually produced high-rated stuff for me, rated my low-rated stuff as high.
Because unfortunately, one reason why readers can’t tell the difference between the AI and human authors is because they don’t have much exposure to the greats. The average person reads something like 2 books a year, and they probably aren't reading Nabokov.
Have a read through the 10 dragon stories where the prompt was "Meeting a dragon" and you'll see what I mean.
https://mark---lawrence.blogspot.com/2023/09/so-is-ai-writin...
IMO both are simply bad and both contain usual telltales in spades (continuity problems, failed or trite metaphors/analogies, semantic failures, overall feeling of 'wtf is even being attempted here').
I'm not so surprised that people struggled with identifying [1] as human - the confounding factor is that this flash story is unpleasantly written, and it's not easy to realize that its failure modes (eg. trying to cram too much in too short a text) are rather human like. And I'm sure the fact that arguably the hardest to digest and rather bad human story opens the poll might somewhat influence the further analyses.
As others in the poll I failed to identify [5] as AI even though in hindsight the telltales are also there. That's because I rather liked it, and as a result it was harder to be vigilant. I also was very undecided on [8]. Finally I scored 6/8, but I wouldn't say it was easy.
Shame that comparing to the previous contest https://mark---lawrence.blogspot.com/2023/09/so-is-ai-writin... is not straightforward. In that one I scored 9/10 while having very easy time (I didn't even finish reading some of them before making up my mind). I also felt completely excused with my only failure, incorrectly identifying as AI the story written in the style of exhaustingly banal fan fiction. But frankly I found almost all the human stories in the previous edition better then the current ones.
In retrospect ChatGPT4 was a terrible writer. ChatGPT5 seems to be an improvement to the admittedly worrying point. Still not impossible to discover though.
However these are my impressions only and it looks maybe I was lucky and I should not generalize it? According to the website people had serious trouble discerning gpt4 writing also 2 years ago. And I'm rather shocked they did. And that they scored some of those banal AI stories positively.
If it's not luck on my part, then maybe discerning AI writing is a skill very different from 'writing' or 'being deeply interested in literature', skills of people who usually frequent this blog?
[1] https://tinhouse.com/the-huntress/ [0] https://gulfcoastmag.org/stories/2015-barthelme-prize-winner...
Example of a long book I put together in about 20 minutes: docs.google[dot]com/document/d/1mA-q1ugWRa6BaOghUTH17kYdjHpaNPIWgoNKg6OvOko/edit?usp=sharing Yes, there are rough edges, but the character arcs and the lore stay coherent from start to finish.
Disclosure: I am building an iOS app that lets anyone spin up a book with any cast and starting situation, in a voice close to a favorite author, and then steer the story while reading. On-demand long-form content is already here. It is just not packaged very conveniently yet.
Aeolun•5mo ago
For me it’s no different from generating code with Claude, except it’s generating prose. Without human direction your result ends up as garbage, but there’s no need to go and actually write all the prose yourself.
And I guess that just like with code, sometimes you have to hand craft something to make it truly good. But that’s probably not true for 80% of the story/code.
antihipocrat•5mo ago
Perhaps I'm a Luddite, or just in the dissonance phase toward enlightenment, but at the moment I don't want to invest in AI fiction. A big part of the experience for me is understanding the author's mind, not just the story being told
add-sub-mul-div•5mo ago
AI content is really exposing how people fall into a group that does go further than the surface text into deeper layers of context/subtext, and a group that doesn't.
Seattle3503•5mo ago
Aeolun•5mo ago
Also, I guess I missed the brunt of your question, though the answer is similar. Most voice works for most characters. There’s only so many ways to say something, but occassionally you have to adjust the sentence or re-prompt the whole thing (the LLM has a tendency to see the best in characters).
exmadscientist•5mo ago
And unless reading is your day job or only hobby, that's a massive, massive corpus of interesting text. (In just one genre! There are more genres!) So on an absolute scale, there is so much fiction to read with more-than-surface-level meaning that I personally just don't understand why anyone would have the least interest in reading AI slop.
(I also don't have any real interest in most Kindle Unlimited works, probably for similar reasons. Though I am quite certain there are diamonds there, I've just not had particularly much time for/good luck at finding them.)
Aeolun•5mo ago
Also, you say you don’t understand why anyone would be interested in the AI slop. But from the article we learn that one is indistinguishable from the other (apparently even to the one professional author that tried)
exmadscientist•5mo ago
CuriouslyC•5mo ago
sram1337•5mo ago
Aeolun•5mo ago
Though I’ll admit I can’t speak to the quality of that except my own stuff (which I’m naturally predisposed to like).
This was my attempt at fully AI generated (though edited by human):
https://www.royalroad.com/fiction/101072/inherited-wounds
strken•5mo ago
I think the AI seems to struggle with consistency of characters and themes, and particularly with character growth over time: it can write touching moments, but these don't fit properly with the character's actions before and then after. It reads a bit like a story written by a hundred professional authors who can skim-read all the previous chapters but are on a strict time limit and don't have access to each other's notes. This makes me wonder if they're just not giving the AI notes on structure and character.
deadbabe•5mo ago