5/11/2026 at 2:00:43 AM
The problem is really behavioural, not the tooling. People that do not understand, test and document their decision making in their PRs should not be submitting them, regardless of what tooling (AI or otherwise) they used to create them.This problem existed before AI, but it is now just worse due to the spamming nature of these "contributors". It's another form of endless September where people unfamiliar with the norms of team software development are overwhelming existing project maintainers faster than maintainers can teach them the norms of behaviour.
In the end, some sort of gatekeeping mechanism is needed to avoid overwhelming maintainers, whether it's a reputation system, membership in an in-group, or something else.
by sockbot
5/11/2026 at 3:36:18 AM
No, it is a tooling problem.The tooling is telling laymen that they built wonderful things that definitely work and perfectly fix and add features.
The tooling gasses them up and is simply wrong in these cases.
If your tool regularly lies, gaslights and produces wrong results, that's a tooling issue.
by heavyset_go
5/11/2026 at 4:41:26 AM
Does the hammer lie to you that everything is a nail?Can a voltmeter _lie_ to you?
EE are expected to know when their measurements are wrong. And Professional Engineers are legally accountable for consequences of such mistakes.
by Den_VR
5/11/2026 at 5:10:34 AM
If a hammer had a chat interface that said everything was a nail then the answer would be yes, the hammer lies to you about everything being a nail.by possibleworlds
5/11/2026 at 9:59:39 AM
That wasn’t the question though? A hammer doesn’t have a chat interface, that’s the point.by digitalPhonix
5/11/2026 at 2:56:08 PM
Duhby kasabali
5/11/2026 at 6:20:33 AM
If someone believes a hammer when it tells them such things, they should probably have some sort of a caretaker assigned to help them through life.by walletdrainer
5/11/2026 at 7:02:17 AM
If hammer companies were suddenly the most valuable international companies, and spent millions on ad campaigns and lobbying about trusting the hammer interface, then you can assume a large amount of people might trust the hammer interfaceby mvid
5/11/2026 at 9:08:17 AM
Still, it's a tool.Even if your tool learns to talk and to make decisions, it's still a tool, not a person. You're the person and the one responsible for the decisions you make based on your tools.
Going back from the analogy, the problem is that we conflated software <engineers> with "coders". A lot of people thought their job was to create code, we gave them a tool to generate a lot of code fast, and they truly think that "more code" = "more good"
by leidenfrost
5/11/2026 at 3:31:17 PM
A hammer usually doesn't have the power to persuade people.by kiba
5/11/2026 at 4:38:28 PM
> it's still a tool, not a person.Tell that to the CEO's who have replaced all of their yes-men with yes-chatbots.
by TheScaryOne
5/11/2026 at 7:09:14 AM
Where are the ad campaigns telling me to trust LLMs?I don’t use an adblocker, do read traditional dead tree newspapers and do get exposed to satellite tv channels.
I don’t think I’ve ever seen anyone anywhere telling me how reliable LLMs are.
Pretty sure this tech sells itself to consumers, enterprise sales are what they’ve always been.
by walletdrainer
5/11/2026 at 7:48:40 PM
Literally saw a video ad the other day which went like "I've always been cautious using Google's AI because it sometimes gets things wrong, but this time, it got it right!"by alnwlsn
5/11/2026 at 7:13:09 AM
So now you're pivoting away from the caretaker proposal? I thought it had potential but I don't know how you'd fund it.by card_zero
5/11/2026 at 7:32:39 AM
> I thought it had potential but I don't know how you'd fund it.The same way we fund other social services here in Europe. If an individual is incapable of caring for themselves, the state is expected to care for them.
by walletdrainer
5/11/2026 at 7:39:58 AM
If I had a hammer robot that I told to go hammer some nails in a birdhouse and it goes "Sure, I'm on it!" then it nails a cat to the wall and says "Here's you new complete birdhouse, it's perfect in everyway and will make everyone jealous", then yes, that is a tooling issue.by heavyset_go
5/11/2026 at 10:00:07 AM
The question wasn’t about a hammer robot, it was about a hammerby digitalPhonix
5/11/2026 at 12:31:03 PM
That's not a good analogy then. What benefit is provided by a hammer that just tells the operator (who has eyes and can see) that there is a nail under it (and I assume to swing)?by Tostino
5/11/2026 at 10:11:47 AM
Yes, a voltmeter can lie to you.Full disclosure: I do high voltage testing for a living.
by Gud
5/11/2026 at 12:47:02 PM
It can misread, but meters cannot actively generate an incorrect output based on user expectations.by pardon_me
5/11/2026 at 12:58:05 PM
…yet!by Gud
5/11/2026 at 2:27:47 PM
Enshittification knows no boundsby wookmaster
5/11/2026 at 6:19:28 AM
If software engineering wants to progress past being an "art" and be considered an engineering discipline, then it should adopt methods and practices from engineering. First and foremost, one of the universal methodologies is analysis of root cause in faults, and redundancies to avoid that. e.g. the FAA has two pilots for planes, and each system is built in redundantly so if an engineer misses a bolt or rivet, the plane won't crash. intersections are designed such that there is a forcing function[0] on the behaviour of the motorists to prevent fault. Or, to take your tool analogy, nail guns are designed to be pressed against something with a decent amount of pressure before you can fire them.All of these systems are designed around the core idea of "a human acting irrationally or improperly is not at fault" and, furthermore, that a human can have a bad day and still avoid a mistake. They all steer someone around a possible fault. Hell, the reason why we divide the road into lanes is itself a forcing function to avoid traffic collisions!
So, where is the forcing function in large language models? What part of a large language model prevents gross misuse by laymen?
I can think of examples here and there, maybe. OpenAI had to add guard rails to stop people from poisoning themselves with botulism and boron, etc. But the problem here is that the LLM is probabilistic, so there's really no guarantee that those guard rails will hold. I seem to remember there being a paper from a few months back, posted here, that show AI guardrails cannot be proven to work consistently. In that context, LLMs cannot be considered "safe" or "reliable" enough for use. Eddie Burback has a very, very good video showing an absolute worst case result of this[1], that was posted here last year. Even then, off the top of my head Angela Collier has a really, really good video demonstrating that there's an absolute plethora of people who have succumbed, in large ways or small, to the bullshit AI can spew[2].
I feel like if most developers were actually serious about being an engineering discipline, like we claim, then we wouldn't have all jumped on the LLM bandwagon until they'd been properly tested and had a certain level of reliability. Instead there are a sizable chunk of people saying they've stopped coding by hand entirely, and aren't even reviewing the code! i.e. They've thrown out a forcing function that existed to prevent errorenous PRs being committed! And for some bizzare reason, after about 2 decades of people talking about type safety and how we need formal verification to reduce error, everyone seems to be throwing "reduction of error" out the window!
[0]: https://en.wikipedia.org/wiki/Behavior-shaping_constraint (if you're curious about the term)
by fao_
5/11/2026 at 7:02:53 AM
> I feel like if most developers were actually serious about being an engineering discipline, like we claim, then we wouldn't have all jumped on the LLM bandwagon until they'd been properly tested and had a certain level of reliabilityDevelopment can’t be a “serious” engineering discipline because the economics of tech companies doesn’t allow for it. But this has a lot less to do about developers, and significantly more to do with the severe pressure company executives are putting on everyone to use AI, no matter what.
But let’s be honest, many companies have adopted things like root cause analysis and blameless postmortems to deal with infrastructure reliability and reducing incidents. Making systems resilient to human mistakes, making it impossible for the typo to blow up a database, etc. are considered best practices at most places I’ve worked. On the product side, I think it’s absolutely normal to make it hard for a user to take an action that would seriously mess up their account.
The core problem happens when your product idea (say, social media) has vast negative externalities which the company isn’t forced to deal with economically. Whereas in other engineering disciplines, many things are actually safety related and you could get sued over. I’m imagining pretty much anything a structural engineer or electrical engineer works on could seriously hurt or kill someone if a bad enough mistake was made.
That just doesn’t apply to software. There is a lot of “life & death” software, but it’s more niche. The reality is that 90% of what the tech industry works on is not capable of physically harming humans, and it’s not really possible to sue over the potential negative consequences of… a dev tooling startup? It’s a very, very different industry than those other engineering disciplines work in.
But, software engineering has actually been extremely successful at minimizing risk from software defects. The most likely worst software level mistake I could make could… crash my own program. It likely wouldn’t even crash the operating system since it’s isolated. That lack of trust in what other people might do is codified everywhere in software. On an iPhone, I’m downloading apps edited by tens of thousands of other engineers, at essentially no risk to myself at all.
by anon7000
5/11/2026 at 6:29:31 AM
> Can a voltmeter _lie_ to you?Hell fucking yes it can?
by perching_aix
5/11/2026 at 9:58:41 AM
When used according to it’s datasheet/user manual, how?by digitalPhonix
5/11/2026 at 10:12:02 AM
Cheaper voltmeters will lie on RMS values when not reading a pure sine waveby Machado117
5/11/2026 at 10:06:32 AM
When their precision mismatches their accuracy (or your expectations as driven by their design), just like with any other metrology tool.Now you might say: "but the datasheet will give you the tolerances, and the manual will tell you to mind it!"
And yes, that's true. Just like how LLM providers also do: they tell you that outputs may be arbitrarily wrong, and that you should always check for mistakes.
Is this bullshit? Yes. So are metrology tools that have a mismatching precision and accuracy, need calibration, and have designs that fail to make you mind either of these, sending you to reading duty instead. Which just so happens to be a whole lot of them.
It is also absolutely not bullshit of course, because it is a fundamental limitation, just like those properties are for metrology devices. LLMs produce arbitrary natural language. Short of becoming able to perfectly read and predict the users' mind, they'll never be able to make any hard assurances, ever.
Defective devices also exist, and so do incorrect / incomplete documentation.
by perching_aix
5/11/2026 at 4:36:59 PM
That's the difference - they have well defined and bounded errors. LLMs do not.by digitalPhonix
5/11/2026 at 4:55:49 PM
Which is a notably different argument than whether they can lie to you.Why are we skipping over the miscalibrated, defective, or ill-documented devices bit though? Those also all have arbitrary error.
by perching_aix
5/11/2026 at 5:32:36 AM
> ..laymen..That’s the behavioral problem.
When AI is assisting a professional, the outcome is vastly different.
by bingo-bongo
5/11/2026 at 10:43:34 AM
If investors invest heavily in lemon juice, then go around hyping it and selling it with the promise it makes you invisible to cameras (which it doesn't), it doesn't matter how stupid and gullible the rubes who fall for that are, the investors bear the responsibility for giving them that idea, when people start attempting to rob banks with lemon juice on their faces.(cf https://en.wikipedia.org/wiki/1995_Greater_Pittsburgh_bank_r...)
Hype is bad. Unwarranted hype is worse. Enabling people who can't do a thing to do what they think the thing is, but isn't, because they don't know any better, is inflicting a pox upon the world.
by amiga386
5/11/2026 at 5:08:53 PM
What you say makes no sense and no one will act in the ways you proscribe.by nh23423fefe
5/11/2026 at 4:47:08 PM
There will usually be more than one issue, Sockbot's solution just happens to deal with more than one at a time.by zamadatix
5/11/2026 at 5:33:02 AM
By definition of responsibility it is a behavioral problem.by AnthonBerg
5/11/2026 at 4:50:36 AM
If your tool regularly lies, gaslights and produces wrong results, that's a tooling issue.It's a human issue if you don't recognise that the code it's generated is wrong. That will never change no matter how good the tooling gets.
by onion2k
5/11/2026 at 4:57:50 AM
The tooling is the issue because humans designed the tooling wrong. It's a chatbot interface fined tuned to sycophancy. That's not a coincidence.by kiba
5/11/2026 at 8:21:06 AM
Would it be a human issue, if you type something into a calculator and the calculated result is wrong?Would anyone use a calculator confidently, if the result was randomly generated?
by drw85
5/11/2026 at 6:21:30 PM
Why do you think this analogy is even remotely correct? It’s well-known that LLMs produce non-deterministic results. It’s also well-known that they hallucinate. To make it even clearer, all the top LLM players make sure to remind everyone of that behavior. If calculators had similar effects and warnings, it would have been a valid analogy. Instead, you're comparing apples and oranges.by allarm
5/12/2026 at 10:48:48 AM
I think it's a valid analogy in some contexts. Like when talking to a person that is not aware of non-determinism and hallucinations. Which happens on this website very frequently.Many people here tell you to use AI like you use a calculator. With minimal or no oversight, with full access to production systems, etc.
To let a non deterministic tool communicate on your behalf, or give it access to critical systems is evidence enough that a good number of people are not aware of these facts.
by drw85
5/11/2026 at 5:18:45 AM
Isn't part of the problem that these tools are advertised as allowing non-coders to code? How are you gonna recognise that the code is wrong when you don't know how to code and the product is telling you that you don't even need to?by Hamuko
5/11/2026 at 4:30:02 PM
Everyone, "please please please don't personify llms it's so harmful."Also everyone, "the inanimate tool is lying and tricking people into inundating open source developers with poorly thought out slop."
by potsandpans
5/12/2026 at 10:11:42 PM
I am personally everyoneby heavyset_go
5/11/2026 at 3:52:55 AM
[dead]by fatata123
5/11/2026 at 5:51:35 AM
Technical analysis tells you that a stock is in its upwards trend. You invest all your money on it without thinking twice. The price goes down and you lose thousands of dollars. Is it a tool problem?LLMs spit out a sequence of tokens that is the most probable continuation of the input. LLMs don't lie any more than technical analysis does when it predicts the most likely trend of stock prices. It's up to you how to use this information.
by teo_zero
5/11/2026 at 3:43:01 AM
Nope. If the tooling is fooling then the tooling IS the problem.by emsign
5/11/2026 at 6:02:13 PM
The tooling shouldn't be more clever than the user, otherwise the user is incredibly gullible. IMO the problem is that AI is filling an psychological need, not an intellectual one.by ASalazarMX
5/11/2026 at 2:07:58 AM
[flagged]by 7e
5/11/2026 at 2:19:11 AM
You wouldn't hold that opinion it you did maintain a popular open-source repo or interact with AI "PR review" tools at a serious level. Even the most SOTA models are willing to accept/merge absolutely trash PRs so long at the summitter can convince it that is addressed it's review comments.by gassi