Need help?
<- Back

Comments (142)

  • raincole
    > Video games stand out as one market where consumers have pushed back effectivelyNo, it's simply untrue. Players only object against AI art assets. And only when they're painfully obvious. No one cares about how the code is written.If you actually read the words used in Steam AI survey you'll know Steam has completely caved in for AI-gen code as well. It's specifically worded like this:> content such as artwork, sound, narrative, localization, etc.No 'code' or 'programming.'If game players are the most anti-AI group then it's crystal clear that LLM coding is inevitable.> This stands in stark contrast to code, which generally doesn't suffer from re-use at all, or may even benefit from it, if it's infrastructure.Yeah, exactly. And LLM help developers save time from writing the same thing that has be done by other developers for a thousand times. I don't know how one can spins this as a bad thing.> Classic procedural generation is noteworthy here as a precedent, which gamers were already familiar with, because by and large it has failed to deliver.Spore is well acclaimed. Minecraft is literally the most sold game ever. The fact one developer fumbled it doesn't make the idea of procedural generation bad. This is a perfect example of that a tool isn't inherently good or bad. It's up to the tool's wielder.
  • noemit
    Many people don't know this, but the Luddites were right. I studied Art History and this particular movement. One of the claims of the Luddites is that quality would go down, because their craft took half a lifetime to master (it was passed down from parent to chile.)I was able to feel wool scarves made in europe from the middle ages. (In museum storage, under the guidance of a curator) They are a fundamentally different product than what is produced in woolen mills. A handmade (in the old traditiona) woolen scarf can be pulled through a ring, because it is so thin and fine. Not so for a modern mill-made scarf.Another interesting thing is that we do not know how they made them so fine. The technique was never recorded or documented in detail, as it was passed down from parent to child. So the knowledge is actually lost forever.Weavers in Kashmir work a similar level of quality, but their wool is different, their needs and techniques are different, so while we still have craftsman that can produce wool by hand, most of the traditions and techniques are lost.Is it a tragedy? I go back and forth. Obviously the heritage fabrics are phenomenal and luxurious. Part of me wishes that the tradition could have been maintained through a luxury sector.Automation is never a 1:1 improvement. It's not just about the speed or process. The process itself changes the product. I don't know where we will net out on software, and I do think the complaints are justified - but the Luddites were also justified. They were *Right*. Their whole argument was that the mills could not product fabric of the same quality. But being right is not enough.I'm already seeing vibe-coded internal tools at an org I consult at saving employees hundreds of hours a month, because a non-technical person was empowered to build their own solution. It was a mess, and I stepped in to help optimize it, but I only optimized it partially, making it faster. I let it be the spaghetti mess it was for the most part - why? because it was making an impact already. The product was succeeding. And it was a fundamentally different product than what internal tools were 10 years ago.
  • simianwords
    What the author and many others find hard to digest is that LLMs are surfacing the reality that most of our work is a small bit of novelty against boiler plate redundant code.Most of what we do is programming is some small novel idea at high level and repeatable boilerplate at low level. A fair question is: why hasn’t the boilerplate been automated as libraries or other abstractions? LLMs are especially good at fuzzy abstracting repeatable code, and it’s simply not possible to get the same result from other manual methods.I empathise because it is distressing to realise that most of value we provide is not in those lines of code but in that small innovation at the higher layer. No developer wants to hear that, they would like to think each lexicon is a creation from their soul.
  • hwers
    Its unfortunate that there’s mode collapse around what the consensus “best way” to use these things are. It’s too bad we didn’t have a period where these things were great teachers but didn’t attempt to write code because in my opinion the ideal way to use them is not by agents mass producing sloppy buggy disorganized code, but to teach you things way faster than the old alternatives, rubber duck, and occasionally write snippets of functions when your brain is too tired or it’s throwaway cli code or some api you’re not familiar with.
  • theshrike79
    > This sort of protectionism is also seen in e.g. controlled-appelation foods like artisanal cheese or cured ham. These require not just traditional manufacturing methods and high-quality ingredients from farm to table, but also a specific geographic origin.Maybe "Artisanal Coding" will be a thing in the future?
  • DavidPiper
    > This stands in stark contrast to code, which generally doesn't suffer from re-use at all ...This is an absolute chef-kiss double-entendre.
  • doodaddy
    There’s a cold reality that we in this profession have yet to accept: nobody cares about our code. Nobody cares whether it’s pretty or clever or elegant. Sometimes, rarely, they care whether it’s maintainable.We are only craftsmen to ourselves and each other. To anyone else we are factory workers producing widgets to sell. Once we accept this then there is little surprise that the factory owners want us using a tool that makes production faster, cheaper. I imagine that watchmakers were similarly dismayed when the automatic lathe was invented and they saw their craft being automated into mediocrity. Like watchmakers we can still produce crafted machines of elegance for the customers who want them. But most customers are just going to want a quartz.
  • Copenjin
    I instantly remembered the page header, I probably visited this site last time 10 years ago or something.
  • plasticeagle
    Acko.net remains the best website on the internet.
  • emsign
    > It's not a co-pilot, it's just on auto-pilot.Love it. Calling it "Copilot" in itself is a lie. Marketing speak to sell you an idea that doesn't exist. The idea is that you are still in control.
  • kombookcha
    What a wonderful read.
  • einr
    This rules. What a good, sensible, sober post.
  • anilgulecha
    >If you ask me, no court should have ever rendered a judgement on whether AI output as a category is legal or copyrightable, because none of it is sourced. The judgement simply cannot be made, and AI output should be treated like a forgery unless and until proven otherwise.Guilty until proven innocent will satisfy the author's LLM-specific point of contention, but it is hardly a good principle.
  • liampulles
    I see a future where I program at work less, which is sad but c'est la vie. I think the challenge of the job will be heralding and managing my own context for larger codebases managed by smaller teams, and finding ways to allow for more experimental/less verified code in prod. And plenty of consulting work for companies which have vibe coded their business and who are left with a totally fucked data model (if not codebase).A Private (system) Investigator. :)
  • Kwpolska
    > Engineers who know their craft can still smell the slop from miles away when reviewing it, despite the "advances" made. It comes in the form of overly repetitive code, unnecessary complexity, and a reluctance to really refactor anything at all, even when it's clearly stale and overdue.I’ve seen reluctance to refactor even 10+-year-old garbage long before LLMs were first made available to the broader public.
  • vladms
    > Whether something is a forgery is innate in the object and the methods used to produce it. It doesn't matter if nobody else ever sees the forged painting, or if it only hangs in a private home. It's a forgery because it's not authentic.On a philosophical level I do not get the discussions about paintings. I love a painting for what it is not for being the first or the only one. An artist that paints something that I can't distinguish from a Van Gogh is a very skillful artist and the painting is very beautiful. Me labeling "authentic" it or not should not affect it's artistic value.For a piece of code you might care about many things: correctness, maintainability, efficiency, etc. I don't care if someone wrote bad (or good) code by hand or uses LLM, it is still bad (or good code). Someone has to take the decision if the code fits the requirements, LLM, or software developer, and this will not go away.> but also a specific geographic origin. There's a good reason for this.Yes, but the "good reason" is more probably the desire of people to have monopolies and not change. Same as with the paintings, if the cheese is 99% the same I don't care if it was made in a region or not. Of course the region is happy because means more revenue for them, but not sure it is good.> To stop the machines from lying, they have to cite their sources properly.I would be curious how can this be applied to a human? Should we also cite all the courses, articles that we have read on a topic when we write code?
  • est
    I won't call that forging, but commission.btw you can make git commits with AI as author and you as commiter. Which makes git blame easier
  • GaryBluto
    I think it says a lot about this opinion piece that the people agreeing with it are posting short comments saying "So true!" and "Great!" whilst the people criticizing it are writing paragraphs of well-spoken criticism.
  • GuestFAUniverse
    And "lazy".Claude makes me mad: even when I ask for small code snippets to be improved, it increasingly starts to comment "what I could improve" in the code I stead of generating the embarrassingly easy code with the improvement itself.If I point it to that by something like "include that yourself", it does a decent job.That's so _L_azy.
  • wilg
    LLMs are pretty cool technology and are useful for programming.
  • barcodehorse
    Lovely lizard machine.
  • anon
    undefined
  • baq
    Lying implies knowing what’s true
  • phendrenad2
    Sigh. Another one standing on the train tracks giving the approaching train a good scolding. First this article tries to equate AI-generated code with "forgery". Please, tell me how you "forge math". Next, it makes a little dig at senior engineers who use LLMs, because they must not realize that "every line of code is a liability". No no, senior engineers realize this, but they are also adept at observing successes and failures and coming up with a mental model for risk. That's part of keeping an application running, otherwise we'd all still be using jQuery and leftPad. We made the jump to react because we recognized that these NEW lines of code were far more valuable than their "liability". Somehow the author decided to store "liability" in a boolean. Oh, was AI involved, or is that a genuine human error..? Next the article makes a tired appeal to the fact that LLMs are trained on open-source code and are therefore "plagiarizing" this code constantly. This is where the train comes around the mountain. So when the AI generates Carmack's Reverse, is it plagiarizing Carmack or the book that he got the idea from? In what percentages? And what do I do with this valuable insight? Send Carmack $0.01 in an envelope for the privilege? In short, I don't know what the author wants, but I hope writing this helped.
  • anon
    undefined
  • DonHopkins
    Pretend Intelligence (PI) — Design Note & TributeA short design note and tribute to Richard Stallman (RMS) and St. IGNUcius for the term Pretend Intelligence (PI) and the ethic behind it: don’t overclaim, don’t over-trust, and don’t let marketing launder accountability.https://github.com/SimHacker/moollm/blob/main/designs/PRETEN...1. What PI IsRichard Stallman proposes the term Pretend Intelligence (PI) for what the industry calls “AI”: systems that pretend to be intelligent and are marketed as worthy of trust. He uses it to push back on hype that asks people to trust these systems with their lives and control.From his January 2026 talk at Georgia Tech (YouTube, event, LibreTech Collective):https://www.youtube.com/watch?v=YDxPJs1EPS4> "So I've come up with the term Pretend Intelligence. We could call it PI. And if we start saying this more often, we might help overcome this marketing hype campaign that wants people to trust those systems, and trust their lives and all their activities to the control of those systems and the big companies that develop and control them." — Richard Stallman, Georgia Tech, 2026-01-23. Source: YouTube (full talk) — "Dr. Richard Stallman @ Georgia Tech - 01-23-2026," Alex Jenkins, CC BY-ND 4.0; transcript in video description.So PI is both a label (call it PI, not AI) and a stance: resist the campaign to make people trust and hand over control to systems and vendors that don’t deserve that trust. In MOOLLM we use the same framing: we find models useful when we don’t overclaim — advisory guidance, not a guarantee (see MOOAM.md §5.3).[...]Richard Stallman critiques AI, connected cars, smartphones, and DRM (slashdot.org) 42 points by MilnerRoute 38 days ago | hide | past | favorite | 10 commentshttps://news.ycombinator.com/item?id=46757411https://news.slashdot.org/story/26/01/25/1930244/richard-sta...Gnu: Words to Avoid: Artificial Intelligence:https://www.gnu.org/philosophy/words-to-avoid.html#Artificia......currently not responding... archive.org link:https://web.archive.org/web/20260303004610/https://www.gnu.o...
  • einpoklum
    > Open source software maintainers have been one of the first to feel the downsides. ... The last thing they needed was to receive slop-coded pull requests from contributors merely looking to cheat their way into having a credible GitHub resumé... As a result, projects have closed down public contributions and dropped their bug bounties...Has this really been people's experience?I develop and maintain several small FOSS projects, some of which are moderately popular (e.g. 90,000-user Thunderbird extension; a library with 850 stars on GitHub). So, I'm no superstar or in the center of attention but also not a tumbleweed. I've not received a single AI-slop pull request, so far.Am I an exception to the rule? Or is this something that only happens for very "fashionable" projects?
  • cindyllm
    [dead]
  • 5o1ecist
    [flagged]
  • chromehearts
    Incredible website
  • Meneth
    That's a lie.