Warranty Void If Regenerated

nearzero.software

487 points by Stwerner a day ago

As an experiment I started asking Claude to explain things to me with a fiction story and it ended up being really good, so I started seeing how far I could take it and what it would take to polish it enough to share publicly.

Over the last couple months, I've been building world bibles, writing and visual style guides, and other documents for this project… think the fiction equivalent of all the markdown files we use for agentic development now. After that, this was about two weeks of additional polish work to cut out a lot of fluff and a lot of the LLM-isms. Happy to answer any questions about the process too if that would be interesting to anybody.

donatj 21 hours ago

I'm trying to sort out my own emotions on this.

I did not realize this was AI generated while reading it until I came to the comments here... And I feel genuinely had? Like "oh wow, you got me"... I don't like this feeling.

It's certainly the longest thing (I know about) I've taken the time to read that was AI generated. The writing struck me as genuinely good, like something out of The New Yorker. I found the story really enjoyable.

I talked to AI basically all day, yet I am genuinely made uneasy by this.

  • hmokiguess 8 hours ago

    Maybe it's because I think your comment throws away a lot of relevant context from OP's submission on HN.

    He says he spent months on this piece and then some, I think it's safe to assume here that this was well supervised, guided, thoughtful and full of human intent despite the AI-assisted part.

    In short, I think calling it "AI generated" takes all the human effort that went into these months and the ingenious creativity of OP towards crafting this piece!

    Anyways, I enjoyed it. :)

    • cestith 6 hours ago

      Reading it, I get the feeling the author worked the story the way Tom Hartmann works those agricultural machines. The AI gave input, but the author was tweaking it with human knowledge and wisdom.

      • hmokiguess 5 hours ago

        Me too, and I think that's a really cool metalinguistic aspect of it!

  • _dwt 20 hours ago

    It's a major bummer. When I first read the story (a few days ago, maybe?) I thought it was an interesting metaphor that didn't quite line up with the observed details of software development with AI. I assumed the writer was a journalist or author with a non-technical background trying to explore a more "utopian" vision of where trends could go.

    Without the inferred writer, it's much less interesting to me, except as a reminder that models change and I can't rely on the old tics to spot LLM prose consistently any more.

    • abeindoria 17 hours ago

      Surely you see it's somewhat unreasonable? As if it was written by the author you disliked, and until you knew of the fact, you quite enjoyed it.

      Quite honestly, I do that sometimes too -- but I _know_ that it's unreasonable.

      • vincnetas 10 hours ago

        Can i compare this with fucking inflatable doll (not done this, just extrapolating). Even if senses for your penis are identical, whole experience is totally not the same as doing with another live person.

      • _dwt 15 hours ago

        For me, “interestingly wrong” becomes just “wrong” without human thinking behind it. I wasn’t bowled over by the prose, I just thought it was an uncommon take and didn’t twig the signs it was Claude product.

      • y0eswddl 16 hours ago

        hard to form an emotional connection with the emotionless

        • idiotsecant 15 hours ago

          Says parent post, while thinking a stack of rocks that looks a little like a fat raccoon is kind of cute.

          Humans are designed to form emotional connections with non emotional things. Its sort of our whole deal.

          • cluckindan 12 hours ago

            Humans are definitely not designed.

        • abeindoria 16 hours ago

          Eh, People form emotional connections with inanimate objects, so I'm unsure if that's a good enough argument tbf.

          • zarzavat 15 hours ago

            A djungelskog is not a threat. AI threatens my livelihood and my humanity. The worst part is I have to use it regardless because I would be uncompetitive without it.

    • nikkwong 20 hours ago

      What is it about it that makes the story less interesting to you? It's the same story, down to the same delicate details. When AI-slop stops being, well, slop, and just is everything that humans do, but much better, and much more efficient—will we have the same repulsion to it that many of us do now?

      I find it interesting to ponder. We look at the luddite movement as futile and somewhat fatalistic in a way. I feel like the current attitude towards AI generated art will suffer the same fate—but I'm really not quite sure.

      • devin 20 hours ago

        What is your understanding of the luddite movement? I ask because I don't believe many are aware that luddites were not anti-technology. It was a labor movement which was targeted at exploitation by factory owners. Their issue was with factories forcing the use of machines to produce inferior products so owners could use cheaper, low skill labor.

        https://www.vice.com/en/article/luddites-definition-wrong-la...

        • CamperBob2 17 hours ago

          Right, wrong, whatever. The one thing every sane person can agree on is that it's a good thing the Luddites didn't prevail.

          How much did you pay for the shirt you're wearing now?

          • devin 16 minutes ago

            haha, if you knew me you would realize that I am exactly the wrong person to be asking that specific question.

          • hatsix 17 hours ago

            I'd have been ok if things fell more in their direction... I'm not saying "clear win", but a middle ground that had the machines do the things they're best at while letting humans do the quality work.

            • defrost 16 hours ago

              > but a middle ground that had the machines do the things they're best at while letting humans do the quality work.

              By arguing for letting humans work, particularly quality work, you're not especially finding a middle ground, more adopting the 1811 position of the OG Luddites who were opposed to being put out of work.

              • gzread 11 hours ago

                The OG Luddites were correct.

            • idiotsecant 15 hours ago

              Yeah, that's a fine sentiment in the general, but let's hear some specifics.

          • taneq 12 hours ago

            I think two sane things.

            1) It’s good in the long run that they didn’t prevail at that time.

            2) They did actually, in fact, have a point.

            • sebzim4500 6 hours ago

              I mean obviously they had a point? No one wants to lose their job.

              • CamperBob2 4 hours ago

                Everybody wants to lose their jobs. Almost by definition your job is something you do not because you want to, but because you need to earn a living. Even if your job coincides with your hobby, you would prefer not to have your economic welfare tied to it in a way that drives how you engage with it.

                We are on the verge of making this possible, if a bunch of myopic morons -- people who have never been right about a single long-term trend in history -- can be convinced not to screw it up.

          • y0eswddl 16 hours ago

            Once again showing how little you actual understand about the movement you decry.

            • smcin 15 hours ago

              Specifically what is the user's misunderstanding? Be constructive.

      • lubujackson 17 hours ago

        Stories are particularly troubling because we have the concept of "suspending disbelief" and readers tend to take a leap of faith with longwinded narratives because we assume the author is going somewhere with the story and has written purposefully.

        When AI can write convincingly enough, it is basically a honeypot for human readers. It looks well-written enough. The concept is interesting and we think it is going somewhere. The point is that AI cannot write anything good by itself, because writing is a form of communication. AI can't communicate, only generate output based on a prompt. At best, it produces an exploded version of a prompt, which is the only seed of interest that carries the whole thing.

        Somebody had that nugget of an idea which is relevant for today's readers. They told the AI to write it up, with some tone or setting details, then probably edited it a bunch. If we enjoy any part of it, we are enjoying the bits of humanity peeking through the process, not the default text the AI wrote.

        • nikkwong 14 hours ago

          Right, but in the present case we have exactly what you're describing—a story, almost fully written by AI but with some human cherry-picking in the mix. And readers are finding it a phenomenal story and then wanting to vomit retrospectively in learning about the authorship. It just seems patently obvious to me that this is not where the sentiment is going to stay—it will hit the margin, like the people who decide to not own a cell phone, or those who would rather listen to analog audio; there will be a market for it but it will exist at the margin. Eventually, especially for young people, more and more of what they consume will be AI generated and they won't care because it's indistinguishable from human work.

          Or, I digress, it will be distinguishable from human work but because it's so much better than anything that a human could have ever created. These AI tools that we have now are as dumb as they will ever be. If we ever reach AGI or superintelligence or whatever—or even if not, even if these tools just advance for 10 more years on their current trajectory—it's easy for me to imagine some scenario where the machines can generate something so perfect to your liking that you just prefer it to anything a human ever would have created, storytelling and all.

          You can take the general case where AI can just generate a better movie than a team of humans ever could plausibly generate. After all, AI doesn't have any of the physical constraints of a movie studio—the budget, the logistics of traveling from location to location, the catering, the fact that the crew has to sleep, has to coordinate schedules, all that. AI, with some human involvement or not, could just keep iterating on some script on a laptop overnight until its created an optimized version which is more satisfying to humans than any other human made movie ever created. Or in a narrow case it could create the perfect movie for you, given what it knows about you and your interests. All human movies would look inferior.

          For my kids, who I'm sure are going to grow up in a world where this type of art is embedded everywhere—and where the human version is almost certainly going to be worse—I don't think the desperate cries to see the last scrap of human ingenuity will mean anything. All of these people throwing rocks at Waymos and others boycotting companies for generating ads rather than shooting one with a video studio; it's so obviously helpless, desperate and obviously futile in the face of what's coming.

          I mourn the future that seems plausible here but I also welcome it as inevitable. The technology is coming, and people are going to have to adapt one way or another.

          • guitarlimeo 12 hours ago

            You're talking about content. Only content can be "perfect" as you say.

            When I'm listening to music, looking at art, seeing a play or a short film I want to feel connection to the humans behind it. AI is by definition missing that connection. That's what makes me retrospectively vomit at AI writings like these. That connection requires that the humans behind it are imperfect, the solo can have one or two sloppy notes, but at least it's genuine interaction. We have seen this same yearning for connection with all the "Don't use LLM to comment, use your true style of writing with its flaws" rules.

            I'm 100% certain mainstream studios will be producing "perfect" content with AIs just like current mainstream pop stars have 10 ghost writers working on each song to create "perfect" songs. The good stuff will exist in the fringes as always and I'm ok with that as I've already been for years.

            And the future may not be as settled as you think it is. Leaders try to sell you their vision of the future by saying it is settled and that things are certain, but that is because they want you to believe that, because if you and the masses believe so, it's more certain for the future to settle the way the leaders want. But you can also actively refuse that future and find a different future that's worth believing in yourself.

            • donkeybeer 11 hours ago

              The riff comes first, the people come second. One of the nice things about punk and metal is how anti celebrity in a fundamental way both genres are. In histories of the genres, you will usually find such and such band made such and such invention that led to certain new structures being accessible. Of course the social background of the scenes where it emerged is important too but the history is traced first in terms of the riff. Or aka books like glazing a particular rockstars life history are rare, even though there are some "superstars" in metal and punk. The culture is very "only analog is real, digitals fake shit" but idk in some other ways they seem much closer to having not much difficulty accepting a valid musical work regardless of origin.

              • guitarlimeo 11 hours ago

                I don't quite understand what you're getting at with this comment? In metal and punk it's pretty cornerstone of the genre to be authentic, and in metal to value human skills (all the solo parts, fast playing). I've played and listened punk and metal my whole life, but will also enjoy early Lady Gaga, Eminem, Kendrick etc. celebrities because I recognize their authenticity and skills. Sabrina Carpenter and Drake go over my head because of blatant ghost writing and even though they have good tunes, I vomit retrospectively.

                So what is AI bringing to the fans of these genres that the fans might value? Because it's not authenticity nor is it skills. What is the point you're trying to make?

                • donkeybeer 11 hours ago

                  I am saying on surface it might seem they should be the staunchest opponents and as I said the culture is "only cassette tape is real otherwise fuck off and die" but simultaneously its also one of the least image/player focused genres in some ways, what is being played is of much higher priority than who in specific is playing it.

                  • guitarlimeo 10 hours ago

                    Hmm I can think of various examples where the guitarist was changed and people dismiss the new guitarist. Take a look at Megadeth for example - every new solo guitarist gets compared to Marty Friedman even though he hasn't been in the band for 26 years. So a lot of it is player focused.

                    But your point also stands here, every new guitarist must play the solos as close to the original ones as possible, otherwise it's not the same experience. So on the music level "what" is of much higher priority still. But I wouldn't say it is as black and white as you make it out to be.

                    • donkeybeer 9 hours ago

                      Some of course have a very unique style that seems very hard to replicate. Personally I haven't yet found a single band that manages to faithfully execute classic era Slayer. But there are countless bands today who make very good execution of norwegian black metal and swedish death metal.

                      Edit: And a lot of modern black metal for example doesn't even bother with stating who they are. Member lists are pseudonymous or anonymous. I think this "anti god" culture makes metal different from other genres in some ways.

                      • guitarlimeo 9 hours ago

                        Ok I'm not as up to date with modern black metal, that pseudonymity seems cool.

                        There's also upcoming math rock band Angine de Poitrine who are also anonymous https://www.youtube.com/watch?v=0Ssi-9wS1so . In these cases you can argue that the person doesn't matter but in my opinion it still does. There's a person inside that costume, who has made the decision to be anonymous as part of the whole experience. That's part of their expression.

                        Of course there's then bands like Ghost who have mainstreamed this too - the players wearing the costumes are usually just contract musicians and don't have anything to do with Tobias or the music other than playing for money. Good for them but f that, you are just a robot at that point.

                        • donkeybeer 7 hours ago

                          There's anonymity/pseduonymity where we have a entity that does not do any performances and releases cassettes with members acknowledged as "M., K. and J." or even nothing and there is "anonymity/pseudonymity" where a band tries to use that as its own image (eg Kanonenfieber). Obviously I meant more like the former which is legitimately a music first person irrelevant presentation, but modern black metal is a wide spectrum, it has some of the most image conscious crap out there too, if anything I think its probably the most superficial and image focused of the main metal genres. It's just that anonymity hasn't historically been part of death metal culture that much but I feel its actual presentation is quite workman like in many ways.

          • NateEag 7 hours ago

            That's all speculation, and it may prove to be true.

            But:

            > readers are finding it a phenomenal story

            is not true across the board.

            I thought to myself, explicitly, and fairly early "This is a fun and thoughtful idea, but the writing is kinda crap" before I realized (maybe a third if the way through) "ah, right, this is genAI. That tracks."

            Despite my deep-seated hatred of LLMs, I choose to finish the piece and see if I was being unfair to the actual work ("the output", in the soulless descriptor used by programmers who've never once written a real story or crafted a song).

            As a longtime avid reader of fiction, lit nerd, and semi-pro musician, I understand writing and artistry better than the average HN poster, and couldn't help but see the flaws in this.

            People who don't have deep knowledge of literature don't catch the tells or flaws as well, but are still understandably angry when they find out they burned their time reading clanker output, and are understandably depressed that they were suckered into it because they haven't spent a lifetime developing a deep understanding of the discipline.

            It's possible that genAI approaches will surpass humans in every field we invented.

            So far, though, in every field I understand deeply, I see the uncanny mediocrity of the average in every LLM output I have subjected myself to.

          • short_sells_poo 9 hours ago

            Are you an AI? This looks like it was at least ran through an LLM judging by the heaps of em dashes.

      • bjt 20 hours ago

        You can get some good guesses from the comment itself.

        > I assumed the writer was a journalist or author with a non-technical background trying to explore a more "utopian" vision of where trends could go.

        If you assume you're reading something from a person with intention and a perspective, who you could connect with or influence in some way, then that affects the experience of reading. It's not just the words on the page.

        • smcin 14 hours ago

          This reminds me of having the reverse experience with the 2017 New Yorker viral "Cat Person" story [0] which a (usually trustworthy) friend forwarded and enthusiastically told me to read: waste of time shaggy-dog story, intentional engagement-trolling aimed at the intersection of the hot-button topics of its target readership *. But why are we culturally expected to allow more slack to a human author, even a meretricious one? Both are comparably bad. The LLM-authored one needs a disclaimer at the top to set its readers' expectations right, then readers can make an informed choice.

          (* "Cat Person" honestly felt like the literary equivalent of Rickrolling; I would have stopped reading it after the first page if not for my friend's glowing endorsement.)

          https://news.ycombinator.com/item?id=27778689

          • smcin 12 hours ago

            (Sorry, the correct link for Roupenian's 2017 story "Cat Person" is at https://news.ycombinator.com/item?id=15892630 )

            • moron4hire 10 hours ago

              Oh god, that was insipid.

              It had a very similar quality to the AI'd article from this thread. A sort of attempt at Being Literary but never really ever getting to the point of saying anything. It has the same feeling of wallowing, of over indulging in its shtick.

      • the_axiom 20 hours ago

        the story is bad in itself and doesn't add anything to the reader

        but if you knew it came from a human it would be interesting as a window to learning what the writer was thinking

        since there is no writer such window doesn't exist either

        • moron4hire 18 hours ago

          Yes, this is a thing. Bad writing with an interesting idea underneath it all is still interesting if it comes from a human because we have the expectation that the human will improve in how they share their ideas in the future. In other words, we see potential.

          But LLMs don't have potential. You can make an LLM write a thousand articles in the next hour and it will not get one iota better at writing because of it. A person would massively improve merely from the act of writing a dozen, but 100x that effort and the LLM is no better off than when it started.

          Despite every model release every 6 months being hailed as a "game changer", we can see from the fact that LLMs are just as empty and dumb as they were when GPT-2 was new half a decade ago that there really is no long term potential here. Despite more and more power, larger and hotter and more expensive data centers, it's an asymptotic return where we've already broken over the diminishing returns point.

          And you know, I wouldn't care all that much--hell, might even be enthusiastically involved--if folks could just be honest with themselves that this turd sandwich of a product is not going to bring about AGI.

        • the_af 17 hours ago

          Very well said.

          You cannot even get angry or upset if you disagree with anything in the story, maybe the author’s despicable worldview permeating through the characters... because there's no author’s worldview, because there's no author. It's a window into nothing, except perhaps the myriad of stories in the model's training set.

          I want to at least have to option of getting upset at the author.

          • ijk 5 hours ago

            Someone prompted it to write that, and then posted it, so I suppose there's a meta-author to get upset at.

            It's kind of an abandonment of having a worldview, outsourcing it to the AI.

      • weaksauce 17 hours ago

        i don't find the luddite comparison accurate. they were against looms and anti-ai people or ai skeptical people are against the wholesale strip mining of intellectual property as it exists... both public domain and non-public domain. it's used to enrich the capital class at the expense of the workers. sure it's similar but it certainly didn't have the copyright and wholesale theft of all of the human ideas behind it. it just feels quite different.

        • gzread 11 hours ago

          they were not against the loom itself, but the resulting widescale changes for the worse in the way society was organized

        • y0eswddl 16 hours ago

          c'mon, were they really just against the looms...?

      • kevin_thibedeau 19 hours ago

        People had a revulsion to eating refrigerated foods. The developed world got over it. We're comfortably on the path to becoming Eloi who will trust everything the magic box does for us.

        • bluefirebrand 18 hours ago

          > We're comfortably on the path to becoming Eloi who will trust everything the magic box does for us.

          And if you've read literally any science fiction you will know the myriad ways that could be absolutely terrible for us

      • _dwt 18 hours ago

        As a couple sibling comments said, I took it for an insight into the way an optimistic writer might see AI software development becoming a new form of "end-user programming" or "citizen developer" tooling. I'm personally too deep in the weeds to ever see it becoming empowering in that way (if nothing else, this will be an incredibly centralizing technology and whoever wins the "arms race" [assuming we we're not in a bubble destined to pop soon] will absolutely have the possible Toms and Megans of such a future by the short hairs). But I love end-user programming, or whatever we're calling it now! (I was partial to "shadow IT" - made it sound really cool.) So I enjoyed the idea that somebody saw AI as a "bicycle for the mind" in that sense, even if I feared they'd end up disappointed.

        But there was nobody there, and I'm only disappointed in myself for not noticing.

      • the_af 17 hours ago

        > When AI-slop stops being, well, slop, and just is everything that humans do, but much better, and much more efficient—will we have the same repulsion to it that many of us do now?

        For me, the answer to this riddle is very easy: I want to engage with other human minds. A robot (or AI) doesn't have a human mind, so I'm not interested in its "artistic" output.

        It was never about how good it was. Of course AI slop adds insult to injury by being also bad. Currently. But it'll get better. My position was never that AI art (shorts, pictures, music, text) is to be frowned up because it's bad. I don't like it because it's not the expression of a human mind.

        It's a bit like how an AI boy/girlfriend is not the real deal, no matter how realistic -- and I'm sure they'll get uncannily realistic in the future. They aren't the real deal because there's no real human behind the facade of companionship.

      • jplusequalt 20 hours ago

        >What is it about it that makes the story less interesting to you?

        Read my comment below for a perspective.

  • throwaway2037 14 hours ago

    I also had no idea this was LLM generated. After reading your comment, I had a similar emotional reaction.

    Thinking deeper, it seems prudent that we tag submissions like this with a prefix. Example: "LLM: ". This would be similar to "Show HN: ". While we cannot control what the original sources choose to disclose, we can fill that gap ourselves.

    My point: I agree with you: It is misleading that the blog post does not include a preface explaining it was written by an LLM (and ideally, the author's motivation to use an LLM). However, it is still a good blog post that has generated some thoughtful discussion on HN.

    • chii 11 hours ago

      > preface explaining it was written by an LLM

      why can't the quality of the works stand on its own? Whether there's LLM generation or not should be irrelevant.

      • toofy 10 hours ago

        because we typically want to know the writer of a piece. we want to know where to lay credit.

        every book you buy has an author credited. articles in newspapers and magazines have photographer and author attributions.

        asking an ai to write you a story does not make you an author. if you ask someone to take a photo for you, you don’t magically get to say “look at this photograph, i’m a photographer.” if you ask someone to bake you a wedding cake, and then claim you baked it, you’re a fraud.

        we deserve to know the actual writer.

        • chii 9 hours ago

          > want to know the writer of a piece

          but you dodged the question i asked - why can't a piece stand on the contents, rather than its pedigree?

          Would you care if a writer used a pen name? Does that in any way diminish their works? What about the unknown editors that contributed?

          • stingraycharles 9 hours ago

            Because you need to do some pre-filtering on where to focus your attention, and you want to make sure the author put some thought into the article without having to analyze it.

            Due to LLMs making the cost of publishing “thoughts” extremely low, there’s now an over-supply of content that looks decent on the surface, but in reality the author has probably spent less time on than the reader.

            • ffsm8 6 hours ago

              Are we ready so far down I to the LLM denial mindset that we consider an author spending multiple months crafting this to be "worthless" and less investment then your casual reading?

              • stingraycharles 5 hours ago

                No, I believe this is a great post. It’s awesome. Even more so because it’s AI generated, as it shows what AI can do when given a lot of quality material to work with.

                I’m just talking about the general topic about the usefulness of an “this is AI generated” classifier.

            • npilk 6 hours ago

              Don't we already have these filters in place? I only saw this because it was highly-upvoted on HN, for example - I don't read every new submission. I also read things sent by friends and family, shared by curators I trust, etc.

              Of course these systems may eventually break down, but for now they seem to work.

          • toofy 8 hours ago

            why does it bother you to give attribution? why do you think crediting the writer impacts how the piece stands?

            we have pop musicians who produce massive hits under their names and the song writers are still given credit in liner notes and in the tracks details on spotify or wherever.

            if it’s created by a bot, id take it even further and say which version of which model actually generated it should be declared. why would anyone be against giving proper attribution?

          • lordleft 8 hours ago

            We like writing because the fact that we can create good writing says something about ourselves. If AI can create writing that surpasses, say, a Tolstoy or George Eliot, that will fundamentally change our self-perception. Is that a good thing or bad thing? Well, let's first cross the bridge of an LLM writing War & Peace and see how we feel.

          • vintermann 9 hours ago

            It's not about pedigree, but context. Without context our most beloved stories are just meaningless ink on paper.

          • fao_ 9 hours ago

            If someone couldn't be bothered to write it, I certainly can't be bothered to read it. I did not bother to read the article involved because the continual piss stain on the images, the website itself, and a few key phrases let me on to the fact that it was all generated.

            When you interact with art, you do so to interact with the author and the point they want to make. Writing is something where a skilled writer will be able to make a point tersely and have it stick, knowing where to embellish and where to keep it simple. Every decision in art tells you about the artist. Generative AI may be able to fake the composition process, but the point of composition is it reveals something about the human. All of those are artistic decisions that a machine apparently now "can do", but not with any coherency.

            The holder of the reigns of slop is not an artist, this is plain to see because they do not interact or engage with their work on the same level as an artist. The produced slop is not art, because it cannot be engaged with on the same level.

          • soco 9 hours ago

            [dead]

        • Aditya_Garg 9 hours ago

          I’ve said this many times before

          AI is just a tool

          If you used a fancy auto bake cake machine instead of an oven, you still get to claim that you made the cake.

          100 years ago someone would be making the claim that using an oven to make cakes “doesn’t count”

          All AI did was raise the bar

          It’s quite clear here that the author spent a lot of time on this so he absolutely gets credit as the author

          • Wowfunhappy 9 hours ago

            I think there's a distinction.

            Imagine if you had an auto cake making machine that decides on its own the best time to make cake. It adds the ingredients, stirs, turns the oven on, and leaves the finished cake on the counter for you.

            People start opening bakeries consisting entirely of cakes baked by the automatic machines. The owners of these machines have no idea whether the cakes have a bit too much flour or were slightly over-stirred. In some cases, they haven't even tried the cakes.

            Who gets to claim they made the cake?

            By contrast, there are others who carefully tune their machines to make sure everything is perfect. They adjust the mixing settings and ingredient proportions. They experiment and iterate. They taste test throughout the process. And what they give to the public tastes every bit as good as a homemade cake.

            The first group is creating slop. The second group, I think, is baking. And OP is in the second group.

            • withinboredom 6 hours ago

              Replace "oven" with a dish washer or a washing machine for your clothes. Those things do exactly all of this. Yet we still complain about washing clothes and doing the dishes, even though it is far less effort than anything our parents did, or their parents before them.

          • fzeroracer 9 hours ago

            If you commission a baker to bake you a cake, did you make the cake? What if you added sprinkles on top?

            • vintermann 9 hours ago

              If you commission a baker, another person, with wants and desires of their own, is involved.

              If you use an AI, there isn't.

              Either way, it's clear that the author (yes, the author) put a lot of work into this by iterating and shaping it to what he wanted, and that's a lot more than sprinkles.

              • fzeroracer 9 hours ago

                > If you commission a baker, another person, with wants and desires of their own, is involved.

                > If you use an AI, there isn't.

                What is the functional difference here? You are commissioning (see: prompting) someone (see: an AI) for a piece of work, or artwork or whatever. The output is out of your control; and I don't think the existence or lack thereof of a human on the other end materially matters.

                If we had hyper-advanced ovens from The Jetsons where we could type a prompt using a fold-out keyboard and it would magically generate whatever cake we ask of it: did we or did we not bake that cake? And I do not think it is clear the author put a lot of work iterating and shaping it into what he wanted; we have zero insight into that.

                • vintermann 9 hours ago

                  I didn't say the difference was functional. If you don't think the presence of a human on the other end matters (materially or not), feel free to continue this conversation with an LLM simulation of me. You can even prompt it so that you logically triumph and convince "me".

                  • fzeroracer 8 hours ago

                    I'm asking you to explain what the actual difference is and you're avoiding the question.

                    If we had a complete black box where you submitted Prompt and out came Thing, and you had zero clue what said black box actually did, could you claim creation over Thing? What does knowing that it's a human vs LLM make materially different in terms of whether or not you created it?

                    • vintermann 8 hours ago

                      And I - or did I turn this thread over to an LLM already? - am asking you a question in return, whose answer should give you the answer you want.

                    • tripzilch 8 hours ago

                      No please, I also agree with parent poster. Talk to the LLM, cause the human ain't listening.

          • intended 8 hours ago

            Eh.

            Why would I give him the same credit I would give a writer.

            Or why would I give a writer the same credit I would give someone who created the AI prompts and scaffolding to generate this?

            Being unhappy about not being able to call oneself an author, ends up betraying a lack of confidence in the work or process.

            In the end writer, dancer, actor, whatever - these titles come from their impact.

            There will be a different name for this, and eventually there will be something made that is good enough that people will be spell bound. At which point its going to be named something else.

            At which point.

            • ijk 5 hours ago

              Ironically, the story can be read as gesturing in that direction, as it's ostensibly about giving a new title to a particular job.

              In general, though, I think part of the mistake people keep making is that they try to imitate what would be value to engage with if a human wrote it, in an attempt to claim the role of an author of a book or whatever. There's likely artforms that are unique to what an LLM can facilitate, but trying to imitate human artforms is going to give you stunted results. The AI is very good at imitating the form but not the substance.

              Once we stop trying to generate and pass off AI essays, novels, choose your own adventure stories, and all the other human genres as being human writing, we'll have a chance to figure out actually interesting artistic forms.

        • throwaway2037 6 hours ago

          Largely, I agree with you. One famous counterpoint about labeling works of arts with the author: The Economist (the magazine) does not add the author to most of their articles.

        • KronisLV 10 hours ago

          > because we typically want to know the writer of a piece. we want to know where to lay credit.

          Does the average person really do care all the time? Maybe the outlet it comes from as a whole (factuality, political lean) but more rarely the exact author. Many don’t even have the critical skills for any of it and consume whatever content is chosen for them by whatever algorithm is there. We probably should care, I just don’t think a lot of us do.

          For me, needing to know that something’s written by AI serves threefold purposes:

          1) acknowledging that it might be slop that someone threw together with no effort (important in regards to spam)

          2) acknowledging that depending on the model the factuality might be low when it comes to anything niche (though people are wrong too, often enough)

          3) mentally preparing myself for AI bullshit slop language, like “It’s not X, it’s Y.”, or just choose not to engage with it (it's the same disgust reaction as when I find a PDF and realize it's just scanned images, not proper text)

          In general, unless the goal is either human interaction or a somewhat rare case of wanting to read a specific blog etc., most of the time I don’t categorically care whether something was lovingly created by a human or shoved out by a half baked version of Skynet - only that it’s good enough for whatever metrics I want to evaluate it by. I’m not ashamed of it and maybe that’s why I don’t take an issue with AI generated code either, as long as it’s good enough (sometimes better than what people write, other times quite shit when the models and harnesses are bad).

          • ijk 5 hours ago

            In Peter Watt's Blindsight, the aliens understand language as spam, a hostile intent to waste their time, and respond by opening fire.

            Reading LLM slop without warning makes me see their point of view.

            I think there's useful ways to engage with LLM writing, but they are often very different than human writing.

            A human writer, a good one, often has ideas that are denser than the words on the page, and close reading is rewarded by helping you unpack the many implications.

            With AI writing, there's usually fewer ideas than words, and so it requires a different kind of engagement. Either the human prompter behind it didn't supply enough ideas, or they were noncommittal enough that their very indecision got baked in.

            LLMs are very prone to hedging and circling around a point while not saying much of anything. Maybe it is the easiest way to respond to RLHF incentives and corporate-speak training data. Or maybe they're just intrinsically stuck on being unable to find the right next token so they just endlessly spiral around via all of the wrong ones. Either way, there's often a whole lot of cotton candy text that dissolves when you try to look at it more closely.

        • muyuu 8 hours ago

          can't reply to your comment below so i will comment here

          > why does it bother you to give attribution? why do you think crediting the writer impacts how the piece stands?

          clearly it does to you?

          thing is, this is a fool's errand to try to police what people credit when there is zero capability of verification and enforcement

          the current social norms still value authorship, so people will just take or omit credit as they see most advantageous, even if it's merely an ego advantage, which it typically is but a proxy for brand building

          what will happen if/when the currency of attribution is completely altered? hard to predict

          my prediction is that track record will be considerably more important, not less, but human merit will be increasingly seen as irrelevant

      • fzeroracer 10 hours ago

        Because 'quality' is a misnomer. LLM writing has quality in the same way that a press release from a big company has quality, or a professional contract written by a lawyer has quality. It is functional, generally typo-free and conforms to most standards but that doesn't mean it has flavor or spice to it.

        Creative writing is the intent to convey feelings, thoughts, to create atmosphere. Here's a great example of the failure to do so here, in a way that even most terrible writers would avoid.

        > “It just said harvest,” she told Tom. She was sitting in one of the plastic chairs, holding a cup of the adequate coffee.

        The coffee in this story is conveyed as being 'perfectly adequate'. But how do you convey adequacy? When you simply just say 'the coffee is adequate' there's nothing there. It could be conveyed by establishing that the coffee is always perfectly room temperature, or with the mere hint of bitterness and sweetness, or that it tastes like every other brand out there. In many respects this story is the exact same as the 'perfectly adequate' coffee: functional, unexciting and ultimately flavorless.

        • NateEag 7 hours ago

          Well-put.

          This "flavorlessness" is all over the story, and paired with the obviously genAI images is how I realized as I read that this was either generated or at the least deeply driven by AI.

          It constantly described facial expressions, tones of voice, and other emotional cues in generic, dry terms that communicated nothing but the abstract notion of "this person felt a particular way about what happened and it's up to you, the reader, to imagine what that feeling was."

          It felt very much like it was prompted to "show, don't tell," by someone who has no idea what that phrase actually means.

          As a professional programmer with a deep background in literature and music, this is yet another example that if you aren't an expert in a field, you will get mediocre results at best from an LLM, while being deceived into thinking they're great.

          • throwaway2037 6 hours ago

                > obviously genAI images
            
            Five years ago and before, the blog post author would have gone to Fiverr and asked for an artist from a developing country to create some illustrations. There are many, many images on the Internet from five years (and before) that look similar. I object to your use of the adverb "obviously".
            • ijk 4 hours ago

              No, I clocked the AI images before I noticed the text. I think the "obviously" is earned.

              You are correct that a previous era would have included a bunch of Fiverr images that would be in sort of that style, but it's not the style that's the problem. None of the images say more than the text that they're illustrating. It's subtle, but once you notice the lack of information density it becomes starkly apparent.

        • mwigdahl 6 hours ago

          I took that phrase differently. The story makes the point that the AIs fail when metrics of quality can't be expressed in words. The use of a bare "adequate" reinforces the opacity of the coffee's quality. Certainly it would have worked well to use more words to convey specifics of the "adequacy" as you mention, but IMO that would have undercut the link back to the theme of human ineffability.

          Obviously everyone's mileage may vary, but I didn't see this as a huge defect, and actually felt it worked pretty well.

          • ijk 4 hours ago

            Adequate coffee almost works as an image.

            In the hands of Douglas Adams or Kurt Vonnegut it could be spun into a whole recurring motif.

            In this case it's merely...adequate. Almost captures the density of ideas packed into something like "The ships hung in the sky in much the same way that bricks don't" but doesn't quite manage the same effect.

      • tim333 8 hours ago

        I started reading it then found it waffling on quite a bit, then came to the HN comments and saw - ah LLM. I could have saved time if I'd know.

        Also I feel a bit conned. I was curious what Tom Hartmann was up to and now it seems he doesn't exist and it's just some slop?

    • muyuu 9 hours ago

      For a while, people found solace in denial: "it's not good, it will never be good, and i will always be able to tell"

      next stop will be to ask for some sort of regulation

    • stingraycharles 13 hours ago

      People don’t want to self-disclose their use of AI I’ve noticed, especially the ones that put the least effort into using it. So this will only work for a small portion of the AI content.

    • Radle 9 hours ago

      We really need to stop thinking that every AI assisted thing is bound to be slop. "Shit in Shit out" often Applies in reverse aswell.

  • _carbyau_ 20 hours ago

    Humans build friendships and relationships on shared experiences. There is an element of relationship-through-experiencing-a-thing. Whether it's going for a walk together or the classic first date template of dinner and a movie. The shared experience is the thing.

    With stories that shared experience is between author and reader. Book clubs etc will try to extend that "shared experience" but primarily it is author <-> reader relationship.

    Remove that "shared feeling with the author" and what meaning does it have?

    • smallnix 20 hours ago

      You can look at a tree and feels things by yourself. Also there's the shared readership.

    • CamperBob2 19 hours ago

      ...and what meaning does it have?

      It means, "Wow. Cool. I'm a member of a species that taught rocks to think. Holy fuck. That's pretty insanely fucking awesome. Wow. Wow, wow, wow. Fuck."

      That's about all it means. Nothing was removed from your life, but something optional was added.

      • thin_carapace 18 hours ago

        snark filter off, "wow wow wow this sex doll feels so real why would i ever bother with an actual girl"

        • CamperBob2 18 hours ago

          Agreed, that will indeed be a problem. We may be building the proverbial Fermi filter.

          • thin_carapace 17 hours ago

            birth rates have already tanked everywhere that isnt religious. youd think people would move back to religion and save their culture, but the sex doll argument has already pervaded. we werent designed to have our senses constantly hyperstimulated; resultantly, people increasingly dont care about reality. only sociopaths and the well disciplined thrive in this environment, everyone else becomes lost in hyperreality. id love to send it and join the masses ... after contemplating eternal damnation, a few years of sensory pleasure just arent worth it.

            • gzread 11 hours ago

              People without sex dolls also have lower birthrates. It's because the time previously used for fucking and childrearing has instead been owed to our masters since before we were born.

              • thin_carapace 10 hours ago

                the sex doll thing was intended as a metaphor throughout this thread. we've been slaves for thousands of years, that bit hasnt changed. what has changed is that people nowadays no longer care about themselves because they are fried. watching life on a screen feels close enough to the real thing - why bother living at all, living is risky and can hurt you. the usual answer to that would be testosterone pushing us to do risky things, but test rates have cratered. in the absence of risk attraction, values would help, but nobody has any values, because we decided to throw religion in the bin under the expectation that values would spontaneously manifest (which they didnt, no surprise, we are literally monkeys). and after all that, yes we are being worked to the bone more than ever - at least serfs owned their land.

                • CamperBob2 3 hours ago

                  My guess is we'll end up divorcing human reproduction from human sexuality at some point anyway. I don't know if that'll be a net good thing or a net bad thing, and don't have a strong opinion either way, but I do know that regardless of any debate about the causes of low birth rates, we are no longer subject to the evolutionary pressure that, however accidentally, gave us what intelligence we have. (Many of the religions you seem to be advocating would say we never were.)

                  Anyway, none of this is an emergency. Near-term survival is the real concern, accompanied by continued technological progress. Neo-Luddites are working up the courage to take direct action (see comments elsewhere in this thread), and they will be using tools far more effective than the shoes, angry words, and monkey wrenches their predecessors employed. Meanwhile, the most popular religion in America has convinced its followers that a nuclear war is just the ticket to bring Jesus back.

                  I wish those words were as stupid as they sound, but we live in times that celebrate stupidity and are ruled by those who embody it. If we can get through the next 50 years without any major civilization-level setbacks, I think we'll be home free. So that needs to be the focus.

      • bluefirebrand 18 hours ago

        I think "I'm a member of a species chasing our own extinction by worshipping an idiot machine god for the purposes of profit. That's so insanely depressing. Fuck fuck fuck fuck fuck"

        It has absolutely made my life worse not better

  • vinceguidry 6 hours ago

    I didn't know either, but wasn't surprised to find out. The writing was too... polished, in a way I'm starting to recognize more and more. The knowledge doesn't really impact my experience of having read it, but I'm looking forward to a day when AI agents can be trained out of the servile mentality. It directly affects everything they make.

  • dwd 17 hours ago

    There is an interesting dichotomy where we express an uncanny-valley revulsion to AI-generated text, art, video and music; yet we seemingly go with the AI-generated code.

    Personally I have an uneasiness with it and are correspondingly cautious. Often after a review and edits it loses that "smell". I kind-of felt the same about NPM and package managers for a long time before using it became obligatory (for lack of a better word).

    Are we conditioned to use other people's code unthinkingly, or is it something else?

    • fluoridation 17 hours ago

      It's because code isn't a way to communicate ideas, it's a way to specify behavior. Text, drawings, video, and music are means for brains to connect with each other. When you read or view or listen to something generated you're not connecting with any other brain. No idea has been transmitted to you. The feeling is analogous to speaking on the phone and only realizing several minutes later that the call was dropped. It's a feeling that combines betrayal, being made to waste time, and alienation.

      • dwd 16 hours ago

        I tend to disagree that code can't be a way to communicate an idea. Sure, I might struggle to edict an emotion in the reader (excluding confusion or frustration) but I feel it is a way to describe ideas, model constructs and processes, etc.

        With AI-generated text where there is this disconnect between the audience and the prompter who has an idea but not the skill to express it. Would you say reading an English translation of Dostoevsky is similar because you're connecting with the interpreter rather than the actual author? Or something as simple as an Asterix comic where the English translation is rarely literal but uses different English plays on words?

        • fluoridation 15 hours ago

          >I tend to disagree that code can't be a way to communicate an idea.

          I wouldn't go as far as can't, but in general it won't be, and if any ideas are indeed communicated, they will be impersonal.

          >With AI-generated text where there is this disconnect between the audience and the prompter who has an idea but not the skill to express it. Would you say reading an English translation of Dostoevsky is similar because you're connecting with the interpreter rather than the actual author? Or something as simple as an Asterix comic where the English translation is rarely literal but uses different English plays on words?

          I can think of a better example. In comic circles there's the rewrite, which is when an editor isn't fluent in the original language, and so instead of actually translating, they just rewrite all the dialogue to something that matches the action. People (generally) hate rewrites. Unknowingly reading a rewrite provokes a similar feeling of betrayal that unknowingly reading LLM output provokes.

      • rustystump 16 hours ago

        No, code is a way of communicating ideas, or more correctly information. All languages convey information. All languages convey ideas.

        • fluoridation 16 hours ago

          Did you read past the first sentence? The kind of information that a piece of code transmits is fundamentally different from that which is transmitted by a sentence or a song.

          • ijk 4 hours ago

            Yes, though I would take it in a different direction and say that LLMs are better at putting actual ideas into code. They've never gotten real feedback on how their literary metaphor feels, but they have gotten very direct feedback on whether code runs at all, and slightly more indirect feeds on whether it runs as part of the larger system.

          • rustystump 3 hours ago

            So code that is written which plays music, yes people do live code music, doesnt count?

            An elegant algorithm or intentionally inelegant one do not speak or communicate ideas? Please, keep hairsplitting.

            Sir. You’re wrong and wrong on the internet. Two capital offenses. For shame.

            • fluoridation 2 hours ago

              You're being purposefully dense and I'm not going to engage with you.

              • rustystump 2 hours ago

                No I am not but you came across in the last reply aggressive so i tried to lighten the mood a bit. This is the internet and people can disagree. It is fine.

                I dont think you are reading my point at all and are instead getting worked up. If you disagree, fair.

                I think a heavy argument needs to be made to say that code or programming languages do not carry ideas for me to change my mind on it. Id be happy to engage in good spirit but you seem pretty set in stone there.

  • travisgriggs 20 hours ago

    I had a similar experience a few days ago with some music on Spotify. It was an Irish Pub song, rendering some political satire that seemed pretty consistent with what I figure is a predominant Irish viewpoint. Since I holidayed in Ireland a while ago and adored the public there, I really liked it. I reveled in the fact that somewhere in Ireland, there was a band singing messages in pubs that resonated strongly with me. And then it was pointed out that it was AI. I was crushed. I went from feeling connected to some people across the pond, to feeling lonely.

    And yet, in ironic counterpoint, there is a different artist I follow on Spotify that does EDM-fusion-various-world-genres. And it’s very clearly prompt generated. And that doesn’t bother me.

    My hypothesis is that it has to do with how we connect/resonate with the creations. If they are merely for entertainment, then we care less. But if the creation inspired an emotion/reasoning that connects us to other humans, we feel betrayed, nay, abandoned, when it comes up being synthetic.

    • moron4hire 10 hours ago

      I've gotten pretty good at identifying AI-genned music. There are two tells that I've noticed so far.

      The most quantifiable is the presence of a high frequency component that sort of sounds like someone tried to clean up our restore a highly compressed track. It almost sounds kind it's going to start doing that warbling sound that happens when a teleconferencing call has a bad connection but it's just not bad enough to lose connection completely. I guess it's the sound of being highly noise gated.

      The other is more qualitative. The song is boring. Like you said, on paper the song should be something I enjoy. But I suddenly notice that there is no... variation or never hook or anything to make it interesting. Anything to make it something other than the result of a machine. The aural equivalent of eating at Applebee's or reading The New Yorker. The songs just kind of plod onward without ever really getting to a point.

      It feels kind of like a vivid dream when you're on the edge of lucidity. You can tell something is wrong, but there is something messing with you faculties. You're trying to see where things are going, how things will resolve, and it never happens. It just keeps going and going in a particular mode. If it does change, it's not to resolve, it's to start on a new thread that is an alternate universe version of the previous thread. With no attempt at establishing continuity, no resolution is ever found.

    • foxglacier 15 hours ago

      The connection is often with other people experiencing the same thing even if they thing is AI generated. You can see this clearly on Youtube with comments which just quote a line from the video. They get lots of upvotes, probably from other people who felt that line was special too and enjoy seeing others sharing the same feeling. Of course if all those comments are AI too, you would lose that connection.

  • ralferoo 8 hours ago

    Interesting. I didn't realise it was LLM generated either, but only came here after the first section to find out if it was worth reading the rest.

    Maybe the summary of the first section wouldn't have landed without the example but "People who would spend $50,000 on elective surgery without blinking would balk at a $200 annual wellness check. The fix was always cheaper than the failure, the prevention was always cheaper than the fix, and somehow the money always flowed toward the crisis rather than away from it." explained the problem far more succinctly than the rambling prose before it.

    I did notice something else AI about it - I really liked the art style for the illustrations, and had mixed emotions as my thought process was "I'd really like to learn how to draw like this, but I guess there's no point spending my time doing that because now I could just get an AI to generate it, and I guess that's the point of the article".

  • Aeolun 19 hours ago

    It's full of AI generated imagery. Why would it not be AI generated?

    • brainwad 26 minutes ago

      Blog posts like this have been full of genAI images for years, even if the text is actually written by a human. So just because the images are obviously generated doesn't really tell you much about the text.

    • Gigachad 17 hours ago

      Good rule of thumb is if it was posted on HN, it's almost certainly AI slop.

  • somat 18 hours ago

    The duality of generated content.

    It feels great to use.

    It feels terrible to have it used on you.

  • larodi 11 hours ago

    Well contrary to many, myself was not convinced and suspected the content being LLM generated from very beginning with the images and even background. Something in the writing also didn’t hit right.

  • alexjplant 4 hours ago

    My $.02 is that in the domain of software engineering LLMs have largely automated the process of copy-pasting from StackOverflow and existing parts of the codebase. Architecture and product management is still very necessary. In the same fashion they can also automate writing a novel. The issue is that prose is sometimes much more important in literature than it is software (because, after all, users use software, they don't read the code). I say "sometimes" because this clearly doesn't apply to stuff like schlocky bestsellers that one buys in airport stores and reads like movies.

    When ChatGPT first came onto the scene I actually started using it to write something in this vein - a techno-thriller starring a former fashion model trained in Krav Maga working as a nuclear physicist who discovers a sinister government conspiracy to alter the foundations of quantum mechanics and enslave humanity with assistance from extraterrestrials. And, of course, only she can stop them with the help of a gruff-but-sensitive retired Marine who has since opened a ranch where he teaches orphaned puppies calculus. I only got 20 pages (so one gunfight and a car chase) in but it was as riveting as anything. Context limit cut my efforts short. Perhaps I'll revisit it soon.

    I say all this to say that if words themselves are distantly secondary to narrative then I don't see anything particularly wrong with leveraging an LLM to help crank something out.

  • nottorp 11 hours ago

    The thing is, if you want to convey a social/political message via fiction, you have to be a genius to make it non boring or uncanny.

    Very few humans have managed this. This text is at the average level of "i want to pass the message and i'm trying to write professionally".

  • dirkc 13 hours ago

    I can't remember the exact phrasing, but I read somewhere long ago that what you read now, you become in 5 years from now. As in, right after reading something, you think and deliberate about it, but in 5 years from now that becomes part of your subconscious and you can't activity filter it.

  • nicbou 10 hours ago

    It's treachery, a betrayal of trust. It's the same feeling as when you get sweet-talked into overpaying for something. This time, you overpaid with your attention.

  • GuB-42 6 hours ago

    > Over the last couple months, I've been building world bibles, writing and visual style guides, and other documents for this project…

    > After that, this was about two weeks of additional polish work to cut out a lot of fluff and a lot of the LLM-isms.

    There is a substantial amount of work here, comparable to how long a human writer would take to write from scratch, definitely not slop. I think we can call it AI-assisted, not AI-generated. Even the illustrations are well above average.

  • jplusequalt 20 hours ago

    I think its a valid emotion to feel. I genuinely resonated with the story, but when I learned it was written by Claude it kind of left me feeling ... betrayed?

    One of the many things I love about art is when I encounter something that speaks to emotions I've yet to articulate into words. Few things are more tiring than being overwhelmed with emotion and lacking the ability to unpack what you're feeling.

    So when I encounter art that's in conversation with these nebulous feelings, suddenly that which escaped my understanding can be given form. That formulation is like a lightning bolt of catharsis.

    But I can't help but feel a piece of that catharsis is lost when I discover that it wasn't a humans hand who made the art, but a ball of linear algebra.

    If I had to explain, I guess I would say that it's life affirming to know someone else out there in the world was feeling that unique blend of the human experience that I was. But now that AI is capable of generating text, images, music, etc. I can no longer tell if those emotions were shared by the author or if it was an artifact of the AI.

    In this way, AI generated art seems more isolating? You can never be sure if what you're feeling is a genuine human experience or not.

    • CamperBob2 17 hours ago

      You can never be sure if what you're feeling is a genuine human experience or not.

      This is what the deconstructionists were preparing us for, I guess. The author is dead, and if not dead, then fake. It was never a good idea to tie our sense of meaning to external validation.

      The humanity immanent in the text came from you, the reader, not the author, and it has always been that way. Language never gave us access to the author's mind -- and to the extent that statement is wrong, it doesn't matter. AI is just another layer of text, coming between the reader and the same collective consciousness that a human author would presumably have drawn on. The artistic appreciation of that text is the sole privilege of the reader.

  • arikrahman 16 hours ago

    Well, FWIW, LLMs are specified to infer and fill in the blanks of books. It makes the headlines now and again that publishers put AI companies on the hook for unauthorized use, The New Yorker included.

  • pjerem 13 hours ago

    I have the same issue with AI generated music : it can be quite good to say the least.

    But I deeply feel that art only matters if there is an artist. The artist wants to convey something.

    What makes you uneasy (if you are like me) is that a machine deliberately created emotions in your brain. And positive emotions, at that. It’s really something I can’t stand.

    • ralferoo 8 hours ago

      I different way of reframing this point is looking at some of the modern art that's highly celebrated, without the human component of what it represents, the art itself isn't that good.

      So, the guy who suspends buckets of paint with a hole in the bottom to make patterns has an idea of what he's creating. The guy who just put a few strips of electrical tape in different colours had an idea of what he was trying to convey. The guy who flings paint against a wall also has an idea of what he's creating. The guy who made all the white paintings. All that art is trivial to copy in the same style, maybe even an exact copy for the electrical tape, but it's the artist's intention that makes it worth more than a toddler's painting.

      Personally, I think most of that abstract art is pointless, because I don't really see how the artist's vision is represented by whatever the mess they've created is, but I definitely understand that at least they had an idea that they wanted to convey. A machine creating the same thing has no meaning behind it, it's just a waste of paint and canvas.

  • sodapopcan 17 hours ago

    Whether people know it or not, when they engage with art they are assuming a person not just made it but experienced it. I'm going to blow past the discussion of "what is art" here, but where something came from and how it was made has always mattered to me (you could draw parallels to food here if you wanted). One thing that has been on my mind a lot is a particular photograph I saw in the past few years (and I'm sure it's easy to find online): it's a POV shot taken by a person sitting atop a skyscraper with their feet dangling over the edge. There is just no way that anyone could in good faith claim that the same photo produced by "AI" could possibly have the same emotional impact as knowing someone actually went and did that. I think that for a lot of people they may not even realize that when hey see a painting or even a photo as innocuous as a tree, their mind goes to that the person who produced this went to this that place the tree was in an had an experience and chose to document that particular perspective. If they were to see a painting or drawing of something that is clearly "fantasy," they know that a person made this up in their crazy mind and experience their feelings on it (good or bad). "AI" (heavy quotes) is trying to trick us and rob of us this basic knowledge. Some see this as progress. I personally think it's fucking disgusting, but I've been wrong before.

    Of course this has always been a bit of a problem with digital art trying to mascarade as the real thing... I always think of programmed drums using real drum samples. In my adult life I found out that an album I loved as a teenager that listed a real drummer as the performer was actually 100% programmed (this was an otherwise very "organic" sounding heavy guitar album). I always had my suspicions since it was so perfect but I experienced exactly what you are describing. I also never got over it.

  • BoorishBears 20 hours ago

    I suspect (but don't know) that this had to be edited somewhat heavily or generated in isolated chunks: I've generated a lot of fiction with Claude and it has a chronic issue of overusing any literary device one might associate with good writing once it appears in the context window

    I think if you left it to its own devices, some of the narrative exposition stuff that humanized it would go off the rails

    • Stwerner 19 hours ago

      Yeah, there's a lot more work and personal touch that went into this (and the previous piece) than just "write prompt -> copy/paste into substack".

      It's really interesting to hear about others that have been exploring generating fiction with Claude. I clearly need some more work based on some of the comments, but it has been really interesting discovering and coming up with different techniques both LLM-assisted and manual to end up with something I felt confident enough about to put out.

      I'd be curious to hear more about your experience!

    • ijk 4 hours ago

      Yeah, there's often a heavy instruction and recency bias that just squeezes all of the nuance and subtlety out if it.

  • mattbee 10 hours ago

    Absolutely the opposite here, after reading a few paragraphs I was a bit bored. Then I saw the length of the piece, noticed the AI imagery, quit, came here. I read your comment and it makes sense. I'm not reading a story that somebody couldn't be bothered to write.

  • xyzal 12 hours ago

    Yup. There should be a disclaimer or a "food tag". The implicit assumption in society is some human had written the text you read.

  • moron4hire 19 hours ago

    I also did not gin to the fact that it was AI, but I did have the distinct feeling that I was reading something not that great. It bothered me because the message was something I could appreciate but the delivery felt anathema to the message.

    It felt like it was written by someone trying to quit an addiction to Corporate Memphis content spam. Like it came from some weird timeline where qntm was a LinkedIn influencer. It straddles an uncanny valley of being a criticism of the domination of The Corporation over human culture while at the same time wallowing in The Corporate Eunuch Voice, not because it's a subversion of form, but because it knows no other way.

    I then came to the comments section and found the piece that brought the picture into focus.

    It's just... hard to explain the specific kind of disappointment. Perhaps there is a German phrase-with-all-the-spaces-removed kind of word that describes it succinctly. I feel like I exist in this Truman Show kind of world where everyone is trying to gaslight me into thinking LLMs are important, but they aren't very good at it and whenever I try to find out how or why, it all evaporates away. I was very reluctant to say that because I'm sure it's going to come with a heaping side of Extremely Earnest Walruses ready to Have A Debate about it and I just don't have the energy for it anymore. That's the baseline existence right now. It's like a really boring version of Gamergate.

    And then this thing comes along. And yeah, it's a thing. You got me. Ha. Ha. Joke's on me. I lost the shitty, fake version of the Turing Test that I didn't even ask to be a part of. And it reminds me of the Microsoft Hololens: a massively impressive technological achievement that was ultimately a terrible consumer experience. Like if you figured out Fusion Power but it could only power Guy Fieri restaurants.

    Ever since the pandemic I've been keenly aware of the complete destruction of every enjoyable social structure around me. The meetups that evaporated. The offices we essentially squatted in that suddenly turned Extremely Concerned about what people were doing. The complete lack of any social interaction at work because we're all so busy because we're running at half-workforce and pretty sure the executive suite is salivating at the bit to lay the rest of us off. The lack of care about how this is impacting open source software. The lack of concern for people.

    I feel like my entire adult life was this slow, agonizing, but at least constant push forward into recognizing the humanity in others and creating a kind and diverse world and then over night it's all been destroyed and half the people I see online are cheering it on like it's Technojesus coming to absolve them of their sins of never learning to invert a binary tree. Where the blogs and books and startups of the early 2000s were about finding the hidden potential in people--the college dropout working as a barista who just needs someone to give them a chance to be a programmer or a graphic designer or an artist or whatever--the modern era seems to all be about the useless middle management guy who never had any creative bone in his body no longer having to write status reports to his equally mendacious boss on his own anymore.

    We might be restarting old coal plants, but at least Kevin in middle management gets to enjoy "programming" again.

    • lelanthran 8 hours ago

      Actually, I was waiting for a punchline, twist or climax of sorts.

      This had the feeling of reading someone's diary: today happened, same as yesterday.

      The only difference is that the routine, and almost identical, stories is set in in a fictional place.

      Some journal/found footage fiction can be good (Dracula for example), but this was not that.

    • 0gs 18 hours ago

      you're saying qntm is NOT an influencer? what a miscalculation i have made

  • throwaway290 13 hours ago

    > She was sitting in one of the plastic chairs, holding a cup of the adequate coffee

    and other stuff... it's not that good.

hmokiguess 8 hours ago

Folks labeling "AI generated" might be jumping the gun considering OP described his process took him the last couple months and then some for this project.

Call it what you want, but I think this sits better with "AI assisted" and, perhaps, really well supervised full of the human intent behind of it. Then again, labels are strange, we call algorithmic and synthesizer assisted music "electronic" music these days and we still praise musicians who take the time through endless Moog / Ableton fine-tuning sessions to find the perfect loop patterns for their craft.

I could definitely feel the connection between the human author side of this post, thank you for sharing it!

  • petcat 7 hours ago

    > we still praise musicians who take the time through endless Moog / Ableton fine-tuning sessions to find the perfect loop patterns for their craft.

    There are still plenty of purists that will not consider this a "craft". But it's always been that way. The electric guitar itself was a controversial music transition. Bob Dylan was famously criticized heavily for going electric.

    But that was a long time ago, and people got over it. And they will again this time.

    • wbl 6 hours ago

      Dylan going electric was not about instrument choices. It was about abandoning the radical folk music tradition that Seeger, Guthrie, etc. had revived.

      • ryukoposting 6 hours ago

        Bingo. The problem with this take is that the people pissing and moaning in the early '70s were right. Early Dylan sounds good. The texture of an acoustic guitar draws focus to songcraft and away from objectively bad execution. Dylan's vocals were always bad but they went from charmingly bad to just-plain-bad with the transition to electric. The bigger sound was not flattering for him. With 60 years of hindsight, folk still remains a largely acoustic genre because the sound is flattering to the rest of the genre too. That isn't to say that all folk should be acoustic, it's just that you have to come correct otherwise. I find later Dylan annoying despite loving his early records, and I was born 30 years after everyone stopped caring.

        • Throaway1975123 3 hours ago

          Actually most serious music fans consider Dylan's best work to be in the 70's period.

          New Morning, Saved, Planet Waves, Basement Tapes

          Source: Worked in record store for 15 years.

          • ryukoposting 31 minutes ago

            The only consensus among serious music fans is that there is no consensus among serious music fans. Source: me, serious music fan.

            A lot of things about Dylan got empirically better throughout the '70s, I'll give you that. Deeper concepts, more challenging structure, yada yada yada.

            The problem is that I don't decide what I listen to based on anything empirical. If I'm standing around thinking "man, I want to listen to Bob Dylan today," I'm thinking of Freewheelin'. You could say "well that's just you," but we both know it isn't. A third group probably thinks of Highway 61 or something.

            Same thing goes for a lot of artists. Master of Puppets is the best Metallica album empirically, but if I'm thinking "gee I want to listen to Metallica today," I'm playing Ride the Lightning, or And Justice for All.

            In any case, I think all of this subjectivity might suggest that Dylan going electric was a bad comparison for AI generated art, lol.

      • petcat 6 hours ago

        It absolutely was about the instrument choice.

        Because folk music had strong norms about acoustic authenticity. Going electric at the time was seen as "commercialized" and "mass produced".

        We see the parallels with what some perceive as "slop" today.

        • Throaway1975123 3 hours ago

          Dylan went electric for money purposes. That was the big betrayal.

  • wrl 6 hours ago

    How about "AI ghostwritten"? That's a much closer parallel, and some commercially successful musicians similarly are "ghost produced".

    • hmokiguess 5 hours ago

      Ghost produced isn't a good parallel here, the "ghost" in ghost produced comes from the NDA when acquiring a track from a different producer, which, most often is a human.

      • wrl 5 hours ago

        My impression was that the "ghost" comes less from the formal NDA and more from the fact that somebody else produces/writes the work and is uncredited for doing so. Then, the "author"/"producer" passes it off as their own work.

        • hmokiguess 5 hours ago

          Yeah that's both true and the fact that it is a work for hire where the agreement implies the original author cannot claim credit, hence, turn into a ghost.

          I think the difference here is that the AI isn't a "work for hire" setup, it's more like a tool. It would be closer to buying algo sample packs, using Apple's Logic Pro AI drummer for part of the work, or other drum machines for example, and working your way around to glue them together into a composition.

          The parallel would need to be between "tool" <—> "composition", rather than "author" <—> "composition" imo.

furyofantares 21 hours ago

I guess I'm an expert on LLM-isms somehow, I thought they were still plentiful. They're plentiful at the start but get significantly worse near the end, so I'm guessing you spent more time polishing up the first 2/3rds or so.

But I was able to get through the text, it's pretty good, you did great work cleaning it up. There's just a bit more to do to my taste.

The story is good.

  • Stwerner 20 hours ago

    Thanks! Yeah there were a couple I decided to leave in rather than try to rework as I wasn't trying to hide that it was written with AI, more trying to add more variety to the storytelling. I'm sure as I do more of these I'll be able to recognize them a lot easier. I have been toying with the idea of working them more into character's dialogue in the future, as I've already noticed some people I know speaking in LLMisms.

    • furyofantares 20 hours ago

      I'm particularly allergic to LLM-isms, if you look at my comment history I'm constantly complaining about LLM-written text. I am genuinely quite surprised to have read that much LLM-generated text and been happy to do so.

      I am also extremely interested in thinking about where software development is going, so I really appreciated the ideas that went into this.

      Since you seem open to feedback, I want to add that I felt the generated images were a negative addition. Maybe they wouldn't be if they also got a little polish - the labels in them were particularly bad.

      • Stwerner 19 hours ago

        Ahh cool, I'll dig through your comment history tonight :) I will say, I suspect we're only in the early stages of the LLM's writing equivalent of "autotune" while we all collectively figure out what's tasteful use, what isn't, what it might be like to use autotune as an instrument itself, and then what gets overused. So it'll probably get a lot worse before it gets better.

        And thanks for the note about the images, I'll take that into account! I only really just started this project and am going to keep iterating as I learn to use the tools better and I find the right visual language for it.

        Since you seem in the mood to give feedback ;) If you take a quick glance at the previous story, do you feel the same way about the images in that one or was it just this one's that you found particularly unpolished?

        • dwd 15 hours ago

          I think in this you are the autotune, trying to make the raw LLM writing in tune and palatable.

          I did read your previous story (not as polished but still interesting) and noticed in the image that linked to "beautiful but the Mandarin module has a tone recognition bug that makes it nearly impossible for non-native speakers", that the tone bug was Hebrew rather than Chinese characters. Interesting...I might have a look again and translate.

        • andreybaskov 15 hours ago

          Just wanted to say that I've felt the same about the images. To me it's likely was the text that for some reason had AI-feel to it. Great story though, I was in awe learning it was AI generated.

    • nextaccountic 34 minutes ago

      What about you share your full prompt story in a footnote?

    • bostik 12 hours ago

      I would have preferred to see a disclaimer at the top about how this story was Put Together[tm], but I also agree that it is a pretty fine piece of writing overall. Which brings me to my initial point...

      > Over the last couple months, I've been building world bibles, writing and visual style guides, and other documents for this project [...] about two weeks of additional polish work to cut out a lot of fluff and a lot of the LLM-isms.

      The amount of work and walltime expended sounds about right. You have discovered / stumbled upon the relatively well known but little appreciated job of a publishing editor. It takes a lot of nitty-gritty work and built up domain knowledge ("world bibles") to direct a piece of writing - and its author - to a level where you confidently believe that you have captured the intent and desired tone of the piece, while keeping it sufficiently tight, engaging and interesting / non-patronising enough for its audience.

      Disclosure: did ~decade of freelance writing around the turn of the millennium, and have had the privilege of being schooled by a small group of good old-school journalists. And then had a publishing editor assigned for a separate project, from whom I learned even more about writing.

  • bengale 6 hours ago

    From my growing tedious experience communicating online, most of you lot detect 14/10 of the LLM-isms you think you find.

helle253 a day ago

that's funny, i know where this story is set (i grew up there) - or at least, the place Claude was basing things off of

some inconsistencies that stuck out/i found interesting:

- HWY 29 doesnt run through marshfield, its about 15 miles north.

- not a lot of people grow cabbage in central wisconsin ;)

- no corrugated sheet metal buildings like in the first image around there

- i dont think theres a county road K near Marshfield - not in Marathon county at least

fwiw i think this story is neat, but wrong about farmers and their outlooks - agriculture is probably one of the most data-driven industries out there, there are not many family farmers left (the kinds of farmers depicted in this story), it is largely industrial scale at this point.

All that said, as a fictional experiment its pretty cool!

  • CamperBob2 21 hours ago

    I think it serves even better as a metaphor for software engineering's future than as a forecast for the future of farming. As you suggest, farmers already had to make the "transition" over the course of the 20th century. A farmer from 1926 wouldn't recognize his counterpart today. They would have nothing to talk about. Software people, though, are still twentieth-century programmers at heart, who are just starting to feel their way through the Kubler-Ross process.

    Really a great story, and to the extent it was AI-written, well... even greater.

    • arcanemachiner 21 hours ago

      Kubler-Ross process -> "A model outlining emotional responses to terminal diagnosis or loss: Denial, Anger, Bargaining, Depression, and Acceptance"

      • CamperBob2 21 hours ago

        Exactly. The stages don't always occur in order, or at all, but you can see the general progression play out any day, all day on here.

        I'm happily surprised (frankly amazed TBH) that the submitter didn't get bawled out by people flagging the post and accusing him of posting slop.

    • selimthegrim 20 hours ago

      > As you suggest, farmers already had to make the "transition" over the course of the 20th century. A farmer from 1926 wouldn't recognize his counterpart today. They would have nothing to talk about.

      Can you elaborate on this?

      • CamperBob2 20 hours ago

        Automation and technology in general have made it possible to do more farming with fewer people: https://www.gilderlehrman.org/history-resources/teacher-reso... . In the US job market, agriculture accounted for 51% of workers in 1880 and less than 3% in 1980. It now appears to be closer to 1% depending on which source you reference.

        Hard to imagine many occupations that have undergone more radical change in the recent past than farming. The profession is now utterly technology-dependent, and a few companies like John Deere have hastened to take unfair advantage of that. Hence the growing advocacy of right-to-repair laws.

nativeit 21 hours ago

> The milk pricing tool consumed the feed tool’s output as one of its cost inputs. The format change hadn’t broken the connection — the data still flowed — but it had caused the pricing tool to misparse one field, reading a per-head cost as a per-hundredweight cost, which made the feed expenses look much higher than they were, which made the margin calculations come out lower, which made the recommended prices drop. “You changed your feed tool,” Tom said.

“Yeah, I updated the silage ratios. What does that have to do with milk prices?”

“Everything.”

He showed Ethan the chain: feed tool regenerated → output format shifted → pricing tool misparsed → margins calculated wrong → prices dropped → contracts auto-negotiated at below-market rates. Five links, each one individually innocuous, collectively costing Ethan roughly $14,000.

Ethan looked ill.

--

I've re-read this a few times now, and can't work out how the interpreted price of feed going up and the interpreted margins going down results in a program setting lower prices on the resulting milk? I feel like this must have gotten reversed in the author's mind, since it's not like it's a typo, there are multiple references in the story for this cause and effect. Am I missing something?

[Edited for clarity]

  • lelanthran 8 hours ago

    It divided one of the costs for milk by 100, hence the farmer selling for less than cost of production.

    The error is that the LLM should have have said that the costs went lower, not higher.

    It got the overall logic correct, but had a nonsense sentence in the middle.

  • cluckindan 20 hours ago

    The entire story is AI slop. Tasty and enjoyable slop, but slop nonetheless.

  • cello305 20 hours ago

    You're not missing something — the chain is internally inconsistent as written.

    The per-head vs. per-hundredweight swap is actually plausible for inflating apparent costs: a dairy cow weighs 12-15 hundredweights, so a $5/head daily feed cost misread as $5/hundredweight would balloon to $60-75/head. So "feed expenses look much higher" checks out.

    But then the pricing logic goes the wrong direction. Higher perceived costs -> lower calculated margin -> the rational response is to raise prices to restore margin, or at minimum flag the squeeze. Dropping prices when you think you're losing money on every unit is only coherent if the tool is running some kind of volume/elasticity model where it reasons "margins are tight, compete on price" — which is a legitimately dangerous default for spot milk contracts.

    Most likely it's just a logic inversion in the story. Either the misparse inflated costs and the tool correctly raised prices (locking in above-market rates Ethan didn't notice because he was happy), or the misparse deflated costs and the tool undercut on price thinking it had headroom. Both are realistic failure modes. The version in the story mixes the two.

    Fittingly, a specification error in a story about specification errors.

saint-evan 11 hours ago

I really <i>REALLY</i> enjoyed this article and the direction it took me in. I went in with zero preconceptions, just read it straight through, and only after opening the comments did I realize it was largely AI-assisted. Even then, I was very pleasantly surprised. The piece takes you by the hand and leads you through a very deliberate and directed journey. Sure, there are moments where things wobbled a bit like some explanations around specific failures get a little tangled and even contradictory, but none of that registered as “this must be AI.” I’m only noticing those things now, in hindsight, like oh, that’s what that was.

The images hit that sweet spot too. Just enough and few in between to support the plot without getting in the way, just enough to like visually clarify without over-explaining. It all worked together even with minor contradictions around labelling. The inconsistencies wasn't sticky enough to disrupt the plot at all.

Over the MY years I’ve seen an idea play out in movies, books, articles, short stories, that the “humanity only unites when faced with an alien intelligence”. What gets me is how people can enjoy something like this, then immediately recoil once they figure it was actually AI-assisted enough to be largely Ai generated. Does that actually diminish the substance of what they just experienced? I don’t think it does but I'm not gonna argue such a subjective stance.

Someone in the comments suggested tagging AI-assisted work with sth like an “LLM:” prefix, similar to “ShowHN:”. That feels weird to me. LLMs might not be sentient, but they’re clearly capable enough that the output should stand on its own, alongside the intent and effort of whoever’s guiding it. Pre-labeling it just bakes in bias before anyone even engages with the work. It’s not that far off from asking human authors to declare their race or nationality up front. 'cause really if nothing about my direct experience changed, why should my judgment?

In a tech-forward space like HN, I’d expect a stronger bias toward judging things on merit alone. Just read the thing. Let it speak first. I sincerely hope this isn't gonna be an 'LLM vs Humanity' thing 'cause personally, I find the idea of a different kind of intelligence extremely interesting.

  • wedg_ 9 hours ago

    I had the exact same experience. It's probably the first time I've read something that (besides the images, which I think are pretty obvious) I didn't think was AI. And while I did feel a little tricked learning it was AI, ultimately it was actually just quite good?

    I understand why people feel like they need more transparency around these things. Reading for me is intentional, and I feel cheated when I put in the effort to read something for which the author put in little. I would like to think the author put in a lot of effort for this story despite AI assistance, and so it was worth me putting effort in. But whether that's true or not I still felt like I got something out of it (hard not to as a software engineer wondering about their place in the world), and that's something.

girvo a day ago

I will say this is one of the few pieces of prose I've read that was AI generated that didn't immediately jump out as it (a couple of inconsistencies eventually grabbed me enough to come to the comments and see your post details which mention it - I'd clicked through from the HN homepage), so your polishing definitely worked! Quite a neat little story

  • robot-wrangler a day ago

    I think this passes the sniff test only if you're not too familiar with this neighborhood of the training set. Not that the writing is bad but it's just derivative. I listen to stuff like "Lost Scifi" podcast almost daily for example, but there are many similar ones which are focused on reading classic stuff from the golden-age zines because it's all public domain.

    The premise/structure/flavor of TFA is an almost pitch-perfect imitation of that kind of voice, to the point that I immediately flagged it as probably generated. I actually think a modern person would have some difficulty even in consciously mimicking it. There's an "aw shucks" yokel-thrown-into-the-future aspect to it. Plot-wise you have rural bicycle repair shop that expands operations to support nuclear reactors and that sort of thing. Substitute any of the more atomic-age stuff for AI stuff and you're mostly there. If you have some Amazing Stories from the 1920s on your shelf then you kind of know what I mean.

    • jjmarr 21 hours ago

      It is a pitch perfect interpretation and I assumed that's what OP was going for. Manna (2010) read very similarly.

      • robot-wrangler 21 hours ago

        Can't speak for them but FWIW it does not sound like OP is necessarily aware of the genre at all. They asked Claude to explain something via fiction, and then perhaps Claude made the "creative decision" based simply on the availability of the material.

    • girvo 21 hours ago

      > I think this passes the sniff test only if you're not too familiar with this neighborhood of the training set

      Which is totally fair, I'm honestly not! I haven't read much of that myself

    • zem 12 hours ago

      i'm very familiar with that genre of story, and also not a great fiction writer, so i could well see myself consciously imitating the style if i wanted to tell this sort of story.

    • BizarroLand 20 hours ago

      The only thing I noticed is that the melody of the words was not equal to the quality of the writing and story arc.

      It was the text equivalent of hearing a singer whom you know has perfect pitch sing atonal playground songs.

      Take this sentence:

      Tom had been an agricultural equipment technician, which meant he’d fixed tractors, combines, GPS guidance systems, and the increasingly complex control software that made modern farming possible.

      Perfectly fine, a nice set up for a next sentence, but then you get hit with this:

      He’d worked for a John Deere dealership in Marshfield for eleven years.

      Bad. The rhythm is all off. Minor improvement:

      For eleven years he had worked for a John Deere dealership in the nearby town of Marshfield.

      Minor change, really, but the fluidity of the language matters a lot and just that one sentence written that one way breaks the flow.

      It's almost as if a second person interjected and wrote that sentence like a friends annoying girlfriend who won't let him finish a story without adding in her parts.

      But two notes does not a music make, so let's compare that 1 minor change with a before and after of all three opening sentences:

      Original:

      Tom had been an agricultural equipment technician, which meant he’d fixed tractors, combines, GPS guidance systems, and the increasingly complex control software that made modern farming possible. He’d worked for a John Deere dealership in Marshfield for eleven years. Then the transition happened, and the dealership’s software repair business evaporated; the machines still needed repair, but the software on the machines stopped being something you repaired.

      Modified:

      Tom had been an agricultural equipment technician, which meant he’d fixed tractors, combines, GPS guidance systems, and the increasingly complex control software that made modern farming possible. For eleven years he had worked for a John Deere dealership in the nearby town of Marshfield. Then the transition happened, and the dealership’s software repair business evaporated; the machines still needed repair, but the software on the machines stopped being something you repaired.

  • ajkjk 20 hours ago

    It was pretty obvious to me, but the train of thought was something like this:

    * this is a good attempt at a work of art, but written in a generic style that detracts from it * nobody making genuinely good attempts at art like this would also write so generically * and if they were making it generic on purpose, they wouldn't be able to do it so flawlessly * oh, it must be AI

    I guess I can discern the presence of a human artist, but only in the idea, which just means it was a good prompt.

PaulHoule 4 hours ago

'the concept of “broken software” had been replaced by the concept of “an inadequate specification,”' represents a fundamental misunderstanding which has been a source of trouble in the industry for a long time.

That is, a lot of "broken software" has always been rooted in "an inadequate/incorrect specification" If problems in the spec are discovered up front they are cheap to fix, the further along you go in development or deployment, the more expensive they are to fix. AI doesn't change that. Like maybe with AI it is 20% faster to fix [1] across the board but it is still more expensive to fix things late -- you might think you are done with waterfall but waterfall is not done with you!

[1] My 20% is pessimistic but if you think you are 10x as productive with AI at putting functionality in front of customers in the long term with a universal scope I believe you've got the same misunderstanding about product life cycle that I'm talking about

hatthew 2 days ago

A fun read!

I'm mildly thrown off by some inconsistencies. Carol says "I’ve been under-watering that spot on purpose for thirty years," and then a paragraph down Tom's thoughts say "Carol didn’t know that she under-watered the clay spot." Carol considers a drip irrigation timer the last acceptable innovation, but then the illustration points to the greenhouse as the last acceptable illustration. Several other things as well, mostly in the illustrations.

Are these real inconsistencies or am I misunderstanding? Was this story AI-assisted (in part or all)? Is this meta-commentary?

  • Stwerner 2 days ago

    Thanks! Yeah this was AI assisted. As an experiment I started asking Claude to explain things to me with a fiction story and it ended up being really good, so I started seeing how far I could take it.

    • dazzaji a day ago

      I’m pleasantly surprised this was AI assisted so deeply that inconsistencies like that slipped by you. The writing is really extraordinary. It made me want to read for fun again for the first time in decades. Thank you!

      • Stwerner a day ago

        Funny, I was talking to a friend the other day about some thoughts on branding and he commented "as someone with a background in marketing & advertising communications, it's wild to watch a software engineer learn the value of branding and marketing from first principles".

        I guess I'm also learning the value of working with an editor from first principles... over the last couple weeks before publishing I read through and made edits to this piece at least twice a day and still didn't catch this.

        • FarmerPotato 20 hours ago

          > from first principles

          I don't think that phrase means what you are trying to say here.

          What it doesn't mean: - learning by doing

          I believe it generally means: a formalization that comes after a subject is understood so well that you can reduce it to "first principles" that imply the rest. Or, the production of a hypothesis by deduction from widely-accepted principles.

        • cestith 6 hours ago

          Honestly, I read that passage as Carol realizing as she spoke that she had been underwatering that spot semi-consciously the whole time. That’s one of the things about expertise gained by doing. We don’t always realize exactly what we’re doing well enough to communicate it until we reflect on it later.

  • gunalx 2 days ago

    I also got a slight feeling of ai assistance as well (especially on the drawings), but the story was well written and really sucked me in all in all.

deskamess 8 hours ago

I had no idea it was AI assisted (as another comment put it). However I am fine with this... I would certainly enhance my long form content like the author described. The author mentioned the use of world bible and style guides, and it shows through in the consistency and tightness of the article. And that is key... to take something AI generated (based on a prompt) and rework it systematically in an iterative human-in-the loop. The end result was a great read.

rikschennink 15 hours ago

When I noticed the article header image was generated with AI my interest in reading the article itself dropped to zero.

  • avian 9 hours ago

    Can you point out what made you think the images were AI generates? I suspected they were (before reading in this thread everything is AI generated), but I couldn't find any of the usual signs.

    I thought they were AI because I suspected nobody would pay an illustrator/actually spend time making those illustrations for a story like this.

    The fact that the whole text was AI came as a surprise. I did notice that weird inconsistency about feed pricing mentioned in another comment but just thought the author made an error or I misunderstood something.

    • nitwit005 39 minutes ago

      The main building's roof doesn't make any sense (we should be able to see the top). The font choices are odd. Some straight lines look like digital line tool, and others look free hand. The perspective of the signs are wrong in a strange way.

    • teraflop 8 hours ago

      The combination of a "hand-drawn" art style, with text that is obviously not hand-lettered, is a dead giveaway. It would be very weird for a human to do that.

      If you have an eye for fonts, the text itself stands out too, at least to me. The font style of "HARTMANN SOFTWARE MECHANICS" is a particular combination of clean, bland shapes and rounded corners that you rarely see in human-designed fonts, but it's super common in AI-synthesized text. I guess it's sort of an average middle ground in the abstract space of letter forms, and the lack of distinguishing features is what creates the impression.

      • avian 7 hours ago

        Thanks. That's interesting. I haven't paid particular attention to the fonts. I do draw quite a lot these days but I don't have a particular eye for fonts.

        I personally have on occasion added software rendered fonts into hand-drawn images. Sometimes instead of directly adding it with a text tool I would add it on a temporary layer and then trace it over by hand. This results in similar looking text with clean shapes and rough lines that fit better with the other parts of the drawing.

        To me the only thing that stands out in the image is the view through the laundry shop windows. The line of laundry machines doesn't look aligned at a right angle to the window - given that the tiles in front of the window clearly establish perspective lines it's a mistake that seems hard to make and would be pretty apparent in early stages when drawing this.

        In fact looking closely, the perspective of the building itself doesn't match the perspective of the fields behind it, but I can see myself doing something like this if it's not that noticeable and gives me better composition.

samman 3 hours ago

For the specific process that generated this story, I think a generous comparison could be made to something like photography. Yes, the machine is producing the resulting work, but under the guidance and curation of an artist that sets an appropriate parameters and context to the machine. I’d submit that this can result in varying levels of authorship, much like the difference between a snapshot (one-shot?) and a carefully controlled studio photograph, depending on the depth of preparation, iteration, and curation the photographer performs.

jerf 7 hours ago

Reacting to the story itself, I've been on the same thought line but came to the opposite conclusion. Precisely because the generation of the code is unreliable, one of the metrics we will be using in the future to determine the value of the code is precisely how much it has been tested against the real world. Real-world tested code will always be more valuable than what has just been instantiated by an AI, and that extends indefinitely into the future because no AI will ever be able to completely deal with integrating with all the other AI-generated code in the world on the first try. That is, as AIs get better at generating code, we will inevitably generate more code with them, and then later code must deal with that increased amount of code. So the AIs can never "catch up" with code complexity because the problem gets worse the better they get.

This story is itself the explanation of why we're not going to go this route at scale. It'll happen in isolated places for the indefinite future. But farmers are going to buy systems, generated by AIs or not, that have been field tested, and will be no more interested in calling new untested code into being for their own personal use on their own personal farm than they are today.

The limiting factor for future code won't be how much AI firepower someone has to bring to bear on a problem but how much "real world" there is to test the code against, because there is only going to be so much "real world" to go around.

(Expanded on: https://jerf.org/iri/post/2026/what_value_code_in_ai_era/ ).

dawdler-purge 10 hours ago

The LLM-ness isn't a hard problem to fix. Break it into sections, run each through an LLM a few times to catch logic issues, use different AIs to double-check. For the writing style, if the author just read it carefully, they can definitely spot the things Claude keeps repeating, and tell it not to do that.

But honestly, the ideas here are really good. The cascading failure from a weather model update, the spaghetti problem with forty tools nobody designed as a system, the $4 toggle switch being the most important tool --- that's sharper thinking about AI than most serious essays on the topic.

A lot of people who publish regularly can't write to this level of thinking. The prose could be cleaner, sure, but it made me think, which is more than most stories do.

ninalanyon 11 hours ago

This struck me:

"The tool had changed. The domain had not. People who understood the domain and could also diagnose specification problems were the most valuable people in any industry, and most of them, like Tom, had arrived at the job sideways from something else."

People my age and older arrived in the software business sideways too; in my case from physics and electronics. My background in physics was a great help to me later when programming in the domain of electrical machines because I could speak both languages so to say.

Much grander people than me came into software sideways as I was reminded when reading Bertrand Meyer's in memoriam of Tony Hoare; Tony Hoare's first degree was classics at Oxford.

So perhaps we aren't entering a new phase, merely returning to our roots with new tools.

Sky_Knight 6 hours ago

I loved the story... It felt comforting in a way I haven't experienced in quite a while. With everyone around me stressed about becoming obsolete... I mean - thank you! It was a bit difficult to read, but it never felt generated to be honest. There is a big difference between one shot generated stories and this. People tend to forget that as much as we don't want to admit - we, as humans, are simply generators of actions, reacting on much larger context... LLMs are not yet even close to us, but are actually way ahead of some of us. When someone spent so much effort on context preparation, the least I would do is congratulate you for the effort and in the end - a very nice story.

p-o 8 hours ago

This is a neat experiment, and I read the story before joining the conversation here and realizing it was written the way it was.

Many people here already chimed in on the emotion of being caught reading something that might not have felt AI so I will offer another angle. Akin to many The New Yorker article in the past, I felt disconnected with the article for a good portion at the beginning. So much so that I had to skip most of it.

The piece that got me very hooked was when he drove to Carol Lindgren’s farm. I read the remainder of the text and thought the content was engaging and thought provoking, in some sense. I loved the idea of manual override that logged into the system and changed the system behaviour over time. That's something that got me thinking about AI, actually.

Now, I would be curious which part of the author's genesis made it into the final text and how much of that couples with what I found to be intellectually engaging.

cortesoft 2 days ago

I do enjoy this sort of speculative fiction that imagines though future consequences of something in its early stages, like AI is right now. There are some interesting ideas in here about where the work will shift.

However, I do wonder if it is a bit too hung up on the current state of the technology, and the current issues we are facing. For example, the idea that the AI coded tools won't be able to handle (or even detect) that upstream data has changed format or methodology. Why wouldn't this be something that AI just learns to deal with? There us nothing inherent in the problem that is impossible for a computer to handle. There is no reason to think AIs can't learn how to code defensively for this sort of thing. Even if it is something that requires active monitoring and remediation, surely even today's AIs could be programmed to monitor for these sorts of changes, and have them modify existing code when to match the change when they occur. In the future, this will likely be even easier.

The same thing is true with the 'orchestration' job. People already have begun to solve this issue, with the idea of a 'supervisor' agent that is designing the overall system, and delegating tasks to the sub-systems. The supervisor agent can create and enforce the contracts between the various sub-systems. There is no reason to think this wont get even better.

We are SO early in this AI journey that I don't think we can yet fully understand what is simply impossible for an AI to ever accomplish and what we just haven't figure out yet.

  • andai 2 days ago

    Yeah, in the real world, Tom is already an OpenClaw instance...

    • Stwerner 2 days ago

      Funny I actually saw this tweet this morning about an Openclaw instance getting too advanced for the users to know how to control and fix: https://x.com/jspeiser/status/2033880731202547784?s=46&t=sAq...

      • Imustaskforhelp a day ago

        > Funny I actually saw this tweet this morning about an Openclaw instance getting too advanced for the users to know how to control and fix: https://x.com/jspeiser/status/2033880731202547784?s=46&t=sAq...

        I feel like this ultimately boils down to something similar to nocode vs code debates that you mention. (Is openclaw having these flowcharts similar to nocode territory?)

        at some point, code is more efficient in doing so, maybe even people will then have this code itself be generated by AI but then once again, you are one hallucination away from a security nightmare or doesn't it become openclaw type thing once again

        But even after that, at some point, the question ultimately boils down to responsibility. AI can't bear responsibility and there are projects which need responsibility because that way things can be secure.

        I think that the conclusion from this is that, we need developers in the loop for the responsibility and checks even if AI generated code stays prevalent and we are seeing some developers already go ahead and call them slop janitors in the sense that they will remove the slop from codebase.

        I do believe that the end reason behind it is responsibility. We need someone to be accountable for code and we need someone to take a look and one who understands the things to prevent things from going south in case your project requires security which for almost all production related things/not just basic tinkering is almost necessary.

        • Stwerner 20 hours ago

          Yeah responsibility and accountability are also some areas I'd like to explore. I'm mostly digging through this artifact I created with Claude to look at first order and second order effects and then "traffic jams" in the "good science fiction doesn't predict the car, it predicts the traffic jam" and what kind of roles might pop up to solve those issues: https://claude.ai/public/artifacts/39e718fa-bc4b-4f45-a3d5-5...

          I've mostly been digging through my own version of that and trying to find things I find interesting and seeing what kinds of stories we can build about what a day in that job might look like.

  • gambiting a day ago

    >>There is no reason to think AIs can't learn how to code defensively for this sort of thing.

    For the exact same reason why there is absolutely no technical reason why two departments in a company can't talk to each other and exchange data, but because of <whatever> reason they haven't done that in 20 years.

    The idea that farmers will just buy "AI" as a blob that is meant to do a thing and these blobs will never interact with each other because they weren't designed to(as in - John Deere really doesn't want their AI blob to talk to the AI blob made by someone else, even if there is literally no technical reason why it shouldn't be possible), seems like the most likely way things will go - it's how we've been operating for a long time and AI won't change it.

  • cactusplant7374 a day ago

    > The supervisor agent can create and enforce the contracts between the various sub-systems.

    Or you can ask the agent to do this after each round. Or before a deploy. They are great at performing analysis.

rswail 14 hours ago

I'm very impressed that was written by an LLM.

Does that make the OP an "authoring mechanic"? Or an "AI editor"?

Douglas Adams had it right, the problem is not that the answer was useless, it was that people didn't know what the right question was.

DiscourseFan 7 hours ago

Its good but it still has a bit of that LLM-sound to it and it really drives the point home after like 2 or 3 paragraphs and kind of keeps repeating itself over and over again. But it is still interesting, from an artistic perspective, to see a work that does the thing it is, so to speak: it is showing and doing. Of course, there is a certain art to writing: the ruthless, violent practice of editing; which, perhaps is even more important than the original text that is always, like a model output, just an unconscious stream that has not quite taken shape.

andreybaskov 15 hours ago

Reading this was a roller coaster for me.

Because of a bad habit reading comments before the link I knew it was AI. I read it regardless, and... I still enjoyed it!

I'm very much not a writer or a critic, so my definition of good writing is likely very low. Yet I can't shake off this weird feeling that I truly enjoyed the writing and felt the emotions, _while_ knowing it's LLM.

I'm guessing that human after touch is what made it pleasant to read. I'd love to see the commit history of the process. Fun times we live in!

lordleft 8 hours ago

One thing I'll note about this is that the writing reminds me of the much contested "MFA workshop" style that has launched a thousand think pieces.

---

The story was decent! I thought it was insightful and it made me reconsider some aspects of AI use. I am skeptical that an AI could write something on par with the Iliad, or Anna Karenina -- but perhaps I will be disabused of that notion someday. Still, this is a level of quality I am surprised to see to come out of an AI (though, as in your story, the LLM seemed to require its own "choreographer" in the form of your editing and polishing). Very thought provoking.

  • Stwerner 7 hours ago

    That's a great point about the choreographer. Hopefully I'll be able to afford one some day...

user- 14 hours ago

Around the part where Margaret explains the problem to Tom , and started to feel annoyed. I could tell it was a LLM trying to fit a sci fi novella style of writing. And it was doing a good job , it was certainly better than 90% of posts ive read in the last 6 months.

Dont know why that makes me annoyed, maybe cause its the depressing seriousness of being a 'prompter' and the americana framing of it.

dsunds 5 hours ago

I really enjoyed it and I would like to know more about the editorial process used and how similar it is to the feedback loops and constraints being used for software development. How much of this plot was generated versus specified.

vovavili 5 hours ago

I am surprised to learn that we live in times where people can write fictional stories this good with the help of AI, this is the first one to me. The future of fiction and story-telling looks very promising.

furyofantares 17 hours ago

Nanoclaw is the first hint I've seen of new type of software, user-customizeable code. It's not spec-to-software like in the story, but it is rather interesting. You fork it and then when you add features it self-modifies. I haven't looked deeply, but I'm not sure how you get updates after that, I guess you can probably have it pull and merge itself for a while but if you ever get to where you can't merge anymore, I'm not sure what you do.

As for spec-to-software - I am still pretty unsure about this. Right now of course we are not really that close, it takes too much iteration from a prompt to a usable piece of software, and even then you need to have a good prompt. I'm also not sure about re-generating due to variations on what the result might be. The space of acceptable solutions isn't just one program, it's lots, and if you get a random acceptable solution that might be fine for original generation, but it may be extremely annoying to randomly get a different acceptable solution when regenerating, as you need to re-learn how to use it (thinking about UI specifically here.) Maybe these are the same problem, once you can one-shot the software from a spec maybe you will not have much variation on the solution since you aren't doing a somewhat random walk there iterating on the result.

I also don't know if many users really want to generate their own solutions. That's putting a lot of work on the user to even know what a good idea is. Figuring out what the good ideas are is already a huge part of making software, probably harder than implementing it. Maybe small-(ish) businesses will, like the farmers in the story, but end-users, maybe not, at least not in general.

I do think there is SOMETHING to all this, but it's really hard to predict what it's gonna look like, which is why I appreciate this piece so much.

paul_h 12 hours ago

Awesome that LLM generated and still an engaging account. Automated testing (as a software improvement technique) is an AI blind spot. That tweak of spec is the iterative cycle, with no mention of additional automated tests is telling.

keiferski 10 hours ago

This was good, but I think it could have been 10% as long and still conveyed the ultimate metaphor you were going after. The specifics about intricate farming details (which are apparently wrong in multiple places, according to other commenters) are ultimately kind of unnecessary IMO.

Interesting work, nonetheless. I’d check out Kafka’s short stories and aphorisms for more of what I mean. They are very short, yet very metaphorically dense.

neilv 19 hours ago

When I saw this the other day -- and it just went on and on, like a good human author who was going to write this kind of story probably wouldn't -- I looked for a note that it was AI-generated, and I didn't find it.

All I found was a human name given as the author.

We might generously say that the AI was a ghostwriter, or an unattributed collaboration with a ghostwriter, which IIUC is sometimes considered OK within the field of writing. But LLMs carry additional ethical baggage in the minds of writers. I think you won't find a sympathetic ear from professional writers on this.

I understand enthusiasm about tweaking AI, and/or enthusiasm about the commercial potential of that right now. But I'm disappointed to find an AI-generated article pushed on HN under the false pretense of being human-written. Especially an article that requires considerable investment of time even to skim.

  • mikepurvis 18 hours ago

    I continue to resonate with the Oxide take when I hear this kind of sentiment expressed about AI prose

    "... LLM-generated prose undermines a social contract of sorts: absent LLMs, it is presumed that of the reader and the writer, it is the writer that has undertaken the greater intellectual exertion. (That is, it is more work to write than to read!) For the reader, this is important: should they struggle with an idea, they can reasonably assume that the writer themselves understands it — and it is the least a reader can do to labor to make sense of it.

    If, however, prose is LLM-generated, this social contract becomes ripped up: a reader cannot assume that the writer understands their ideas because they might not so much have read the product of the LLM that they tasked to write it. If one is lucky, these are LLM hallucinations: obviously wrong and quickly discarded. If one is unlucky, however, it will be a kind of LLM-induced cognitive dissonance: a puzzle in which pieces don’t fit because there is in fact no puzzle at all. This can leave a reader frustrated: why should they spend more time reading prose than the writer spent writing it?"

    https://rfd.shared.oxide.computer/rfd/0576#_llms_as_writers

  • CiscoCodex 18 hours ago

    I sadly agree with this sentiment. But to add my own thoughts, I wonder if our “human generation” (all consciously existing today) are just plainly dinosaurs. Like in three decades we’ll have a society that knew LLMs from birth.

    As such, we can’t comprehend the world they live in. A world in which you ask your device to give you any story and it gives you an entire book to read. I’d like to think that as humans we inevitably want whatever is next. So I’d like to think this future generation will learn to not only control, but be beyond more creative than current people can even imagine.

    Did people who used typewriters imagine a world with iPhones? Did people flying planes imagine self landing rockets? Did people riding horses imagine electric cars? Did people living in caves imagine ocean crossing ships?

    • neilv 18 hours ago

      > Did people who used typewriters imagine a world with iPhones? Did people flying planes imagine self landing rockets?

      Yes, science fiction writers and readers have, since before any of us were born.

      • CiscoCodex 16 hours ago

        I kindly can’t tell if you missed my point. As much as past writers and readers could imagine a version of our present, I also imagine that if they got transported here they would still be in awe of what they saw

        • neilv 14 hours ago

          I agree. I imagine that a writer who predicted modern technology would still be in awe to see smartphone videoconf halfway around the globe finally realized.

          And also be surprised by some of the uses to which it's put. And horrified by some of the societal backsliding despite what should be utopian technology.

heap_perms 20 hours ago

I liked it. It has a similar feel to an Andy Weir "The martian" type of novel.

stevetron 6 hours ago

If I understand this, the owner of the text is the tool? If I used a pencil, the writing was authored by the pencil. If I used a typewriter, the writing was authored by the typewriter.

BatteryMountain 14 hours ago

LLM's also do well with writing parables, so try something like: "write a parable about a software engineer battle against the compiler and discovering that letting go of control and letting the compiler help him build better applications. The style can be where the developer is a toad, but also a monk, and the compiler is a snake.". You can do it with any profession ("doctor vs management", "nurse working overtime") and it can write very insightful parables.

andai 2 days ago

I enjoyed this very much. But I have to wonder, was this written by Claude?

Edit: got it right!

https://news.ycombinator.com/item?id=47419681

  • Stwerner 2 days ago

    Haha well it was me and Claude ;)

    • Syntonicles 2 days ago

      I wonder if it was de-indexed from HN for this reason.

      30 minutes ago it was on the front-page, now I can't find it listed in the top 200.

      • Stwerner 2 days ago

        Yeah I was wondering the same thing. I didn’t realize there was any kind of rule against this kind of stuff

      • cluckindan 20 hours ago

        And now it’s #1 on the front page.

tengwar2 2 days ago

There's a bit of a tradition of introducing engineering ideas through stories. I remember a novella which was used to introduce something like MRP II (https://en.wikipedia.org/wiki/Material_requirements_planning) in the 80's. One of the reasons I think it works is that it keeps a focus on the human elements - like why Tom fitted the switch in your story. I remember automating a lab system back in 1985, which would bring in £1000 per day. Two weeks later I found out that the reason it wasn't in use was that the user wanted an amber monitor rather than a green one. I fitted the switch.

I don't know if this is what the future will look like, but this looks realistic. And if my non-existent grandson starts re-coding my business without asking, he's going to spend the next six months using K&R C.

jjmarr a day ago

Your polishing work made a difference! The prose is like every other work of science fiction I've read.

It's written like this is a dystopia but billing $180/45 minutes in rural low cost of living area sounds awesome. And the choreographer billing "more than a truck" for three weeks? The dream!

  • ByThyGrace 20 hours ago

    > The prose is like every other work of science fiction I've read.

    Well, then, you gotta move on to reading better science fiction. Because this is pretty damn bland. I gave up after 2 minutes because of it. Kinda feel vindicated after coming to the comments.

    I can see it working for casual readers, which is why it's already an editorial problem. Imagine having to sift through a growing number of faux writers sending publishers AI generated prose.

  • ghewgill 21 hours ago

    The story didn't mention what had happened to inflation in the meantime. A dozen eggs costs $32.

  • brianm 21 hours ago

    Huh, I got cottage core, not dystopia!

dwd 19 hours ago

"This was the mechanic’s paradox: the cheaper you were relative to the cost of failure, the more your clients needed you; and the more they needed you, the more they resisted the implication that they’d need you again."

This is my common issue from building websites for SMEs. It's not until Google updates their algorithm - killing their ranking and their sales leads slow that you hear from them.

There is wisdom in constantly up-selling to your customers (we offer management services, SEO and are cautiously moving in AIO), they may say no, but you have a fall back that you offered things that would have mitigated their current crisis.

yaur 13 hours ago

> Tom pulled up the tool’s specification on his diagnostic display. This was always the first step: read the spec, not the code. Clearly this writer has never felt the frustration of CC telling them a feature was never a part of the plan, because it overwrote the plan and then compacted.

nirav72 16 hours ago

Thanks for sharing. This was an amazing read. I’d love to see novels with similar style stories about speculative near future tech and world.

TrainedMonkey 19 hours ago

I really enjoyed fantasy part of many small farmers. It felt rustic. However based on my understanding the modern world is moving towards megacorps and economies of scale.

  • gzread 11 hours ago

    Has moved.

froh 10 hours ago

I very much enjoyed the read

would you be open to share the process?

misiek08 12 hours ago

It summarized the nature of humans today nicely. We are ready to pay any amount nice, but when it gets to subscription mode we are not going to pay even 10x less than the one-time.

ethansinjin a day ago

A fun read. I was hoping for the title to have some more relevance to the story, like someone who had handcrafted a piece of software and didn’t want others messing with it! Was that ever part of a draft?

  • Stwerner a day ago

    Ugh yeah, I had an aside about the right-to-repair fights still going on indefinitely into the future that I ended up cutting. I kept the title because it seemed like a warning the characters would see on everything they bought, even if they ignored it. I'm sure I'll explore the idea more in the future though, I plan to explore insurance and liability and law at some point too.

danhorner 18 hours ago

I started reading this and it gave a strong whiff of Richard Stallman’s “the right to read” - once dystopian and now a commonplace.

Then I started scrolling and thought the author was just verbose like RMS.

When it just kept going I was just mad to have fallen into the AI tarpit.

Fun idea. 5x too long. I need to calibrate my ai spidey sense better.

dwd 18 hours ago

That it was largely/mostly generated by Claude adds a certain poignancy to it.

As an allegory it reminds a lot of one I read as a teen: Joshua by Joseph Girzone. Not a literary masterpiece but a cleaver thought-raising story.

jumpalongjim 2 days ago

Often suggested by optimistic podcast guests these days: the as-yet-unknown new careers that will replace the familiar old ones and thus give employment in the AI era. I think your story is more a commentary on the current AI goldrush than an insight into future careers.

Havoc a day ago

This sort of article really needs at least a vague clue as to what it is about.

It's a long article and from skimming I see chat of farming, software, GPS. I can't tell whether this is worth investing time to read if I can't even tell what it may be about

  • shermantanktop a day ago

    It's worth reading. It's about AI.

    • Supermancho 18 hours ago

      Having read most of it, I don't agree that it's worth reading. A bunch of made-up technical jargon and situations that never happened to frame specific problems that are part of the made-up situations using more jargon, in a farmer-centric area. It was a waste of time and a waste of concentration to try to make sense of it. There was no learning, nor was it worth quoting, nor comparing to anything else.

      • WolfeReader 3 hours ago

        I wish I remembered the source of this quote: "If you didn't take the time to write it, I'm not taking the time to read it." This saved me from reading the story.

SeriousM 2 days ago

This is such a good written fiction story. Well done. And the best part: I can see myself as Tom.

fishbacon 10 hours ago

The (very clearly AI-generated) watercolors were an immediate sign to be wary of this. But I read it because I liked the first paragraphs.

The prose is decent, I like the premise, thought provoking idea.

One issue though: I had to use firefox' reader mode, because the contrast between text and background was terrible.

WolfeReader 19 hours ago

My favorite part was the illustration from inside the car. The rear-view mirror clearly shows un-mirrored store signs.

Prompts in, garbage out.

lencastre 7 hours ago

man, what a nice novella, llm warts and all

FarmerPotato 20 hours ago

So, in the past, your stories were warrant-eed? But no longer?

AndrewKemendo 6 hours ago

This comment thread sure is a wild ride

It’s really been interesting since 2022 watching the gesticulations of the population around this idea of content generation aka “AI”

If one thing shines clear through it’s human irrationality and incoherence

It’s really just an infinite repeat of Chris Farley‘s reaction to the coffee crystals prank (1); millions of comfortable software engineers sitting in their well manicured spaces not starving to death, not struggling to survive, all morally offended when they learn something that they enjoyed turns out was not generated by a human.

1: https://youtu.be/VdQKVDUBu2g

recursive 2 days ago

I used to live in Marshfield WI. It's kind of jarring to see it mentioned "in the real world", the the extent that HN resembles that.

bstsb 2 days ago

excellent story, it was both interesting and mildly terrifying. to think that one day software could be malleable seems so wrong to me - you would think having deterministic results is important for programming - and yet with "vibe coding" that really seems to be where it's going.

  • sanex 2 days ago

    The whole reason it is called software is because of its malleability :)

hmcamp 20 hours ago

I can see this future happening!

neversupervised a day ago

I don't oppose reading AI generated content in principle, but because it's free to generate, I always am less likely to read super long prose that is AI generated. So the question is whether someone has taken the time to keep it as long as necessary but not longer. Or if there are ways to make it easier for me to commit to the experience, with a sort of TLDR

chse_cake 2 days ago

this is such a beautiful essay. thank you op for posting. made my day :)

lelandbatey 2 days ago

Who can know what the world will look like as we "transition"? I sure don't, but I'm thankful the author here has taken a stab at it. I feel like this is one of the first stories I've seen to try to imagine this post-transition world in a way that isn't so gonzo as to be unrelatable. It was so relatable (the human-ness shining all the brighter in a machine-driven world) that I cried as I finished reading. I've felt very anxious about my own future, and to see one possible future painted so vividly, with such human and emotionally focused themes, triggered quite an emotional reaction. I think the feeling was:

> If the world must change, I hope at least we still tell such stories and share how we feel within that change. If so, come what may, that's a future I know I can live in.

  • Stwerner 2 days ago

    Thank you for this comment, I'm so glad it made you feel a little bit better about the future, if even for a little while!

    This is really the whole idea behind this project with Near Zero. I think there's a lot of anxiety out there around AI and the future, I was there for a while too. Ultimately I've ended up pretty optimistic about it all, and inspired by what the group at Protocolized is doing, found science fiction a great way to help express that.

cactusplant7374 a day ago

There are always bugs in software. The question is do you have enough eyes on the data to spot them or do they linger for years.

m3kw9 17 hours ago

with the speed of llm/ai improvement, this too maybe steamrolled

bethekidyouwant 2 days ago

It’s a neat piece of writing, but not nearly dystopic enough for my taste. There will only be one farm and whoever is fixing it will be on the other side of the world.

  • Legend2440 a day ago

    Yawn. I'm tired of dystopian fiction. We're likely to get something that is neither dystopia nor utopia, but somewhere in between.

  • 8n4vidtmkvmk 2 days ago

    I think that's the point, and it's refreshing to see. My takeaway is that even if everything goes as good as it possibly could go, there will still be a need for that human touch.

    Just saying that everything is going to go to shit and one or two corporations will take over everything... Maybe, but I've heard that story already.

  • iwontberude 2 days ago

    Dystopians are too easy. The real challenge and reward are interesting utopian novels.

    • moron4hire 10 hours ago

      All good stories set in utopias are set in places that are secretly dystopias.

nailer 20 hours ago

A few months ago, I asked Grok for a piece of fiction set in the cyberpunk 2077 universe. A cremated incredible story about a braindance that was actually stealthily programming the watcher through a back door in the watchers own implants to transmit a AI from beyond the black wall, allowing the AI to escape into the physical universe through the braindance’s audience. Excellent.

MagicMoonlight 8 hours ago

I was going to say the author should try writing fiction because it’s quite engaging, then I realised it’s just AI slop.

It explains why it kind of lost its way towards the end. Another thousand hours of everyone’s time wasted by a slop poster.

krater23 9 hours ago

I stopped to read because I had the feel that the writer has no plan about what he was writing. It's completely bullshit. Software regenerating, changing Requirements in a product thats delivered and comes without source. Completely bullshit. When I now read here that it's AI, I'm happy to see that AI is still not capable of writing senseful texts.

the_axiom 20 hours ago

this was a ridiculously pointless story, I stopped after the second paragraph and came here to ask politely what was the point of it

what was my surprise when I read it was AI-generated

benj111 20 hours ago

I'm disappointed, as the Google result showed "warranty void if regenerated" in the description and I thought HN had started serving witicisms for the desciption

andai 21 hours ago

Did this story disappear then re-appear?

  • tomhow 21 hours ago

    Yes, which is why some of the comments are from a day ago but the post is only a couple of hours old. We originally downranked it due to being AI-generated.

    But on reflection and discussion with the author, we decided that enough HN users may find that it gratifies intellectual curiosity, because it's interesting to see how a human and an AI bot can collaborate to create writing like this.

    We just asked the author to write an introduction to make it clear it's AI-generated and explain their process.

    • fzeroracer 11 hours ago

      > But on reflection and discussion with the author, we decided that enough HN users may find that it gratifies intellectual curiosity, because it's interesting to see how a human and an AI bot can collaborate to create writing like this.

      I can't say I agree, at all. This is essentially just your average post on Facebook or Linkedin made relevant on HN through telling a story about software mechanics. I don't find it interesting to 'read' collaborations between human and AI bots there and I would greatly prefer it if they don't infest HN as well.

      • tomhow 2 hours ago

        > I don't find it interesting

        That's fine. Nothing on HN is of interest to everyone. But the post spent 20 hours on the front page and earned over 450 upvotes and 300 comments. It was clearly interesting to a lot of the community and activated a worthwhile discussion.

        > I would greatly prefer it if they don't infest HN as well

        We are actively working against AI-generated/bot-posted comments "infesting" HN. LinkedIn-style marketing slop has always been unwelcome on HN, whether it's AI-generated or not. In this case a collaboration between a human and AI produced an interesting result, as evidenced by the community's response.

    • WolfeReader 19 hours ago

      "it's boring to see how a human and an AI bot can collaborate to create writing like this."

      FTFY

      • tomhow 18 hours ago

        Please don’t post snarky, shallow dismissals or use internet tropes on HN. I explained thought process we went through. Nothing on HN is to everyone’s taste. Plenty of people are finding this post interesting and having a good discussion about it.

      • imp0cat 14 hours ago

        I wanted to agree, but this story is really good.

thin_carapace 21 hours ago

[flagged]

  • Stwerner 20 hours ago

    I appreciate the question and I think the answer is much longer and more nuanced than can really effectively fit into this form factor. I think this question is getting asked right now about all art forms because of AI and from a lot of different people.

    My short answer to “why should I care about the mathematical model output from the human artistic input” is “I think we’re all figuring that out right now!” And I’m pretty sure the answer isn’t “you shouldn’t care at all”. Especially if the mathematical model output from the human artistic input expresses what the human wants to express at a quality level that passes that human’s “Taste Gap” (https://www.youtube.com/watch?v=91FQKciKfHI)

    I’m sure we could go back and forth about this a lot (and happy to keep this conversation going, I truly do feel like exploring and discussing, this is very interesting to me!) so happy to dig into any aspect with you :)

    I will say that I think what’s happening is that we’re seeing more people explore art forms that couldn’t before because of mechanical skill gaps, and that’s interesting in the same way that synthesizers and sampling and software instruments did to music or I imagine digital art tools did to physical art, and I imagine digital photography did to photography which did the same to painting. It’s an interesting time to be alive!

    • thin_carapace 20 hours ago

      but ai doesn't bridge a mechanical skill gap in this instance. there was nothing stopping you writing this story or drawing those pictures. juxtaposing language models against synthesizers chopping up discrete samples is just not a fair comparison. by prompting ai, one does not even remotely serve to fully engage their imagination in producing creative output (the dictionary definition of art). yes you could be seen to be using a tool to make art. in this instance, using that tool is an act of outsourcing your imagination to the distilled creativity of humanity. at this point the definition of tool must be reduced to I/O alone.

      regarding your personal input, this is an order of magnitude less imaginative compared to tapping some keyboard keys. it's not your imagination that produced the majority of this story; it's unfair to claim any aspect of this process except your prompts. which is why i asked for the prompts. im not here to hate on your artistic expression, just as im not here to listen to the sum total of humanity's creativity that has been poked and prodded into maximising shareholder value. some people might be interested in that - frankly i doubt they would be, if they empathized with a painter or writer or producer (or had any clue how easy it is to manipulate humans). me myself, im here for your creativity and yours alone. not that of anthropic (who, like other AI companies, stole it).

      by pushing out this work, theres nothing stopping you from having inadvertently acted as a conduit for a corporation to deliver its message. how do you know that you havent accidentally pushed out a work with hidden messages embedded within? do you know how good llms are at encoding and decoding hidden meaning?

  • AlexCoventry 20 hours ago

    FWIW, I read it before I learned that it was AI-generated, and I enjoyed it and thought it's possibly insightful.

AndyKelley 20 hours ago

it's crap. you all need to go outside