4/10/2026 at 8:00:56 PM
Basically the rules are that you can use AI, but you take full responsibility for your commits and code must satisfy the license.That's... refreshingly normal? Surely something most people acting in good faith can get behind.
by qsort
4/11/2026 at 2:20:42 AM
I agree this is very sane and boring. What is insane is that they have to state this in the first place.I am not against AI coding in general. But there are too many people "contributing" AI generated code to open source projects even when they can't understand what's going on in their code just so they can say in their resumes that they contributed to a big open source project once. And when the maintainer call them out they just blame it on the AI coding tools they are using as if they are not opening PRs under their own names. I can't blame any open source maintainer for being at least a little sceptical when it comes to AI generated contributions.
by pibaker
4/11/2026 at 5:37:02 AM
I think them stating this very simple policy should also be read as them explicitly not making a more restrictive policy, as some kernel maintainers were proposing.by theptip
4/11/2026 at 6:22:53 AM
It cannot be understated how religiously opposed many in the Linux community are to even a single AI assisted commit landing in the kernel no matter how well reviewed.Plenty see Torvalds as a traitor for this policy and will never contribute again if any clearly labeled AI generated code is actually allowed to merge.
by lrvick
4/11/2026 at 8:23:49 AM
Just remember that "reviewed" is not enough to not be considered public domain.It needs to be modified by a human. No amount of prompting counts, and you can only copyright the modified parts.
Any license on "100% vibecoded" projects can be safely ignored.
I expect litigations in a few years where people argue about how much they can steal and relicense "since it was vibecoded anyway".
by Luker88
4/11/2026 at 8:44:07 AM
For those who might wonder how accurate this is, there is advice from the Federal Register to this effect. [0] Its quite comprehensive, and covers pretty much every question that might be asked about "What about...?"> In these cases, copyright will only protect the human-authored aspects of the work, which are “independent of” and do “not affect” the copyright status of the AI-generated material itself.
[0] https://www.federalregister.gov/documents/2023/03/16/2023-05...
by shakna
4/11/2026 at 9:11:01 AM
I cannot take seriously any politician or layer using the words "artificial intelligence", especially to models from 2023. These people have never used LLMs to write code. They'd know even current models need constant babysitting or they produce unmaintainable mess, calling anything from 2023 AI is a joke. As the AI proponents keep saying, you have to try the latest model, so anything 2 years old is irrelevant.There's really 2 ways to argue this:
- Either AI exists and then it's something new and the laws protecting human creativity and work clearly could not have taken it into account and need to be updated.
- Or AI doesn't exist, LLMs are nothing more than lossily compressed models violating the licenses of the training data, their probabilistically decompressed output is violating the licenses as well and the LLM companies and anyone using them will be punished.
by martin-t
4/11/2026 at 10:28:11 AM
If monkeys can't hold copyright, which is an actual case discussed above, then no, an LLM probably can't either. "Human" is required.by shakna
4/11/2026 at 8:35:29 AM
Meanwhile I expect that intellectual property protections for software are completely unenforceable and effectively useless now. If something does not exist as MIT, an LLM will create it.The playing field is level now, and corpo moats no longer exist. I happily take that trade.
by lrvick
4/11/2026 at 8:48:44 AM
Isn't the "corpo moat" bigger now?They can wash the copyright by AI training, but the AIs don't get trained on closed source.
"corpo" also has a ton of patents, which still can't be AI-washed.
What will become unenforceable are Open Source Licenses exclusively, how does that make it a "level field"?
by Luker88
4/11/2026 at 9:13:48 AM
Because AI is also proving to be very good at reverse engineering proprietary binaries or just straight up cloning software from test suites or user interfaces. Cuts both ways.by lrvick
4/11/2026 at 9:19:36 AM
Have you ever seen what obfuscation looks like when somebody puts the effort in?Not to mention companies will try to mandate hardware decryption keys so the binary is encrypted and your AI never even gets to analyze the code which actually runs.
It's not sci-fi, it's a natural extension of DRM.
by martin-t
4/11/2026 at 10:40:56 AM
Companies have been encrypting code to HSMs for decades. Never stopped humans from reverse engineering so it certainly will not stop AI aided by humans able to connect a Bus Pirate on the right board traces. Anything that executes on the CPU can be dumped with enough effort, and once dumped it can be decompiled.by lrvick
4/11/2026 at 10:22:52 AM
[dead]by greton7
4/11/2026 at 9:17:47 AM
Exactly.AI proponents completely ignore the disparity of resources available to an individual and a corporation. If I and a company of 1000 people create the same product and compete for customers, the company's version will win. Every single time. Or maybe at least 1000:1 if you're an optimist.
They have access to more money for advertising, they have an already established network of existing customers, they have legal and marketing experts on payroll. Or just look at Microsoft, they don't even need advertising, they just install their product by default and nobody will even hear about mine.
Not to mention as you said, the training advances only goes from open source to closed source, not the other way around.
AI proponents who talk about "democratization" are nuts, it would be laughable if it wasn't so sad.
by martin-t
4/11/2026 at 8:51:22 AM
The corporate moat is the army of lawyers they have. It doesn’t matter whether they win or not if you can’t afford endless litigation. Is the same for patents.by adrianN
4/11/2026 at 10:15:39 AM
Funny, their army of lawyers seems incapable of stopping me from easily downloading pirated software or coding an open alternative to their closed-source software with AI if I wanted to..You cannot keep a purely legally-enforced moat in the face of advancing technology.
by Marha01
4/11/2026 at 9:17:53 AM
The music industry has an army of lawyers too, and it did not make a damn bit of difference once bittorrent was popularized.IP law means nothing once tens of millions of people are openly violating it.
The software industry is about to learn this lesson too.
by lrvick
4/11/2026 at 10:04:16 AM
So is music free now? The record industry doesn't exist anymore, isn't ridiculously profitable? Artists are finally earning a fair share?by dwedge
4/11/2026 at 10:18:08 AM
> So is music free now?Uhm... yes? The cost of downloading pirated music is essentially zero. The only reason why people use services like Spotify is because it's extremely cheap while being a bit more convenient. But jack up the price and the masses will move to sail the sea again.
by Marha01
4/11/2026 at 10:36:30 AM
The cost of stealing has always been essentially zero. Same argument can be made for streaming, and yet Netflix is neither cheap nor struggling for subscribers.by dwedge
4/11/2026 at 9:19:47 AM
> Any license on "100% vibecoded" projects can be safely ignored.As far as I know that has only been decided in US so far, which is far from the whole world.
by VorpalWay
4/11/2026 at 9:28:44 AM
In Poland law is similar in this regard, so I'd assume at least some other countries do this as well.by IsTom
4/11/2026 at 10:27:08 AM
So, how are you gonna prove I didn't write some code?How am I gonna prove I did?
by OtomotO
4/11/2026 at 9:04:19 AM
I don't think modified by a human is enough. If you take licensed text (code or otherwise) and manually replace every word with a synonym, it does not remove the license. If you manually change every loop into a map/filter, it does not remove the license. I don't think any amount of mechanical transformation, regardless if done by a human or machine erases it.There's a threshold where you modify it enough, it is no longer recognizable as being a modification of the original and you might get away with it, unless you confess what process you used to create it.
This is different to learning from the original and then building something equivalent from scratch using only your memory without constantly looking back and forth between your copy and the original.
This is how some companies do "clear room reimplementations" - one team looks at the original and writes a spec, another team which has never seen the original code implements an entirely standalone version.
And of course there are people who claim this can be automated now[0]. This one is satire (read the blog) but it is possible if the law is interpreted the way LLM companies work and there are reports the website works as advertised by people who were willing to spend money to test it.
[0]: https://malus.sh/
by martin-t
4/11/2026 at 7:17:42 AM
Some people are just against change, that's nothing new. If Linus was like them, he would never have started linux in the first place.by cinntaile
4/11/2026 at 8:37:03 AM
Not every change is good, and sometimes we realise too lateby sdevonoes
4/11/2026 at 8:49:33 AM
What is it that worries you about the change that is happening?by cinntaile
4/11/2026 at 7:59:43 AM
Are they against change in general, or certain kinds of change? Remember when social media was seen as near universal good kind of progress? Not so much now.by goatlover
4/11/2026 at 8:15:56 AM
Social media has never been seen as a universal positive force? It's the same with AI. It has good and bad aspects as does any technology that has an impact on this scale, AI will arguably have a much bigger impact imo.People are generally against change that forces them to change the way they used to do things. I'm sure most will have their reasons why they are against this particular change, but I don't think it will affect anything. The genie is out of the bottle, AI is here to stay. You either adapt or you will slowly wither away.
by cinntaile
4/11/2026 at 10:11:37 AM
It reminds me of something I read on mastodon: "genie doesn't go back in the bottle say AI promoters while the industry spends a trillion dollars a year to try to keep the genie out of the bottle"by dwedge
4/11/2026 at 10:19:51 AM
Do you think the genie will go back in the bottle and why?by cinntaile
4/11/2026 at 10:17:19 AM
Adapting implies you are still a part of the environment though. AI is on a trajectory to replace you and take you out of the environment.by gnz11
4/11/2026 at 8:11:53 AM
This is like blaming a knife as being a killer weapon. Social media is inherently good if owners of the platforms allow for good interactions to take place. But given the mismatch between incentives alignment, we don't have nice things.by contraposit
4/11/2026 at 10:12:33 AM
Social media is good if owners allow for good is an example of the logical fallacy "begging the question"by dwedge
4/11/2026 at 7:35:10 AM
Sounds dramatic, but it entirely depends on what "many" and "plenty" means in your comment, and who exactly is included. So far, what you wrote can be seen as an expectable level of drama surrounding such projects.by dxdm
4/11/2026 at 7:14:14 AM
True - on Mastodon there is a very vocal crowd that are against AI in general, and are identifying Linux distros that have AI generated code with the view of boycotting it.by ebbi
4/11/2026 at 8:31:28 AM
Soon they will have to boycott all of them. Then what I wonder?by lrvick
4/11/2026 at 9:58:28 AM
Doesn't matter. Linux today is a toy of corporations and stopped being community oriented a long time ago. Community orientation I think these days only exists among the BSD and some fringe linux distributions.The linux foundation itself, is just one big, woke, leftist mess, with CV-stuffers from corporations in every significant position.
by abc123abc123
4/11/2026 at 10:19:38 AM
The idea that something can simultaneously be "woke [and] leftist" and somehow still defined by its attachments to corporations is a baffling expression of how detached from reality the US political discourse is.The rest of the world looks on in wonder at both sides of this.
by simonask
4/10/2026 at 8:09:34 PM
But then if AI output is not under GNU General Public License, how can it become so just because a Linux-developer adds it to the code-base?by galaxyLogic
4/10/2026 at 9:28:08 PM
AIs are not human and therefore their output is a human authored contribution and only human authored things are covered by copyright. The work might hypothetically infringe on other people's copyright. But such an infringement does not happen until a human decides to create and distribute a work that somehow integrates that generated code or text.The solution documented here seems very pragmatic. You as a contributor simply state that you are making the contribution and that you are not infringing on other people's work with that contribution under the GPLv2. And you document the fact that you used AI for transparency reasons.
There is a lot of legal murkiness around how training data is handled, and the output of the models. Or even the models themselves. Is something that in no way or shape resembles a copyrighted work (i.e. a model) actually distributing that work? The legal arguments here will probably take a long time to settle but it seems the fair use concept offers a way out here. You might create potentially infringing work with a model that may or may not be covered by fair use. But that would be your decision.
For small contributions to the Linux kernel it would be hard to argue that a passing resemblance of say a for loop in the contribution to some for loop in somebody else's code base would be anything else than coincidence or fair use.
by jillesvangurp
4/11/2026 at 7:25:29 AM
Copyright Office's interpretation of US copyright laws says that AI is not human, thus not an attributable author for copyright registration, and output based on mere prompting is no one's IP, it can't be copyrighted[1].When AI output can be copyrighted is when copyrighted elements are expressed in it, like if you put copyrighted content in a prompt and it is expressed in the output, or the output is transformed substantially with human creativity in arrangement, form, composition, etc.
[1] https://newsroom.loc.gov/news/copyright-office-releases-part...
by heavyset_go
4/10/2026 at 10:41:03 PM
That you can't copyright the AI's output (in the US, at least), doesn't imply it doesn't contain copyrighted material. If you generate an image of a Disney character, Disney still owns the copyright to that character.by nitwit005
4/11/2026 at 4:57:06 AM
> That you can't copyright the AI's output (in the US, at least),It's also not really clear if you can or cannot copyright AI output. The case that everyone cites didn't even reach the point where courts had to rule on that. The human in that case decided to file the copyright for an AI, and the courts ruled that according to the existing laws copyright must be filed by a person/human/whatever.
So we don't yet have caselaw where someone used AIgen and claimed the output as written by them.
by NitpickLawyer
4/11/2026 at 6:28:48 AM
You can copyright AI output assuming there is a "reasonable" degree of human involvement. https://www.cnet.com/tech/services-and-software/this-company...by metalcrow
4/11/2026 at 3:10:57 AM
Yes. And that’s why the rules say that the human submitting the code is responsible for preventing this case.by fxtentacle
4/11/2026 at 5:38:19 AM
> Is something that in no way or shape resembles a copyrighted work (i.e. a model) actually distributing that work?Does a digitally encoded version resemble a copyrighted work in some shape or form? </snark>
Where is this hangup on models being something entirely different than an encoding coming from? Given enough prodding they can reproduce training data verbatim or close to that. Okay, given enough prodding notepad can do that too, so uncertainty is understandable.
This is one of the big reasons companies are putting effort into the so called "safety": when the legal battles are eventually fought, they would have an argument that they made their best so that the amount of prodding required to extract any information potentially putting them under liability is too great to matter.
by friendzis
4/10/2026 at 10:15:58 PM
IANAL; this is what my limited understanding of the matter is. With that caveat: it is easy to forget that copyright is on output- verbatim or exact reproductions and derivatives of a covered work are already covered under copyright.So if the AI outputs Starry Night or Starry Night in different color theme, that's likely infringement without permission from van Gogh, who would have recourse against someone, either the user or the AI provider.
But a starry-night style picture of an aquarium might not be infringing at all.
>For small contributions to the Linux kernel it would be hard to argue that a passing resemblance of say a for loop in the contribution to some for loop in somebody else's code base would be anything else than coincidence or fair use.
I would argue that if it was a verbatim reproduction of a copyrighted piece of software, that would likely be infringing. But if it was similar only in style, with different function names and structure, probably not infringing.
Folks will argue that some things might be too small to do any different, for example a tiny snippet like python print("hello") or 1+1=2 or a for loop in your example. In that case it's too lacking in original expression to qualify for copyright protection anyway.
by ninjagoo
4/10/2026 at 10:40:57 PM
>AIs are not human and therefore their output is a human authored contribution and only human authored things are covered by copyright.That is a non sequitur. Also, I'm not sure if copyright applies to humans, or persons (not that I have encountered particularly creative corporations, but Taranaki Maunga has been known for large scale decorative works)
by Lerc
4/11/2026 at 2:39:03 AM
Copyright applies to legal persons, that's why corporations can have copyright at all.by Sharlin
4/11/2026 at 5:39:36 AM
A "large scale decorative work" is the strangest euphemism for a dormant volcano I've ever heard.by direwolf20
4/11/2026 at 9:00:31 AM
Well obviously it's not doing any decorating right at the moment.by Lerc
4/10/2026 at 11:08:05 PM
Didn't a court in the US declare that AI generated content cannot be copyrighted? I think that could be a problem for AI generated code. Fine for projects with an MIT/BSD license I suppose, but GPL relies on copyright.However, if the code has been slightly changed by a human, it can be copyrighted again. I think.
by mcv
4/10/2026 at 11:33:15 PM
Thaler v. Perlmutter said that an AI system cannot be listed as the sole author of a work - copyright requires a human author.US Copyright Office guidance in 2023 said work created with the help of AI can be registered as long as there is "sufficient human creative input". I don't believe that has ever been qualified with respect to code, but my instinct is that the way most people use coding agents (especially for something like kernel development) would qualify.
by simonw
4/11/2026 at 2:36:05 AM
Interesting. That seems to suggest that one would need to retain the prompts in order to pursue copyright claims if a defendant can cast enough doubt on human authorship.Though I guess such a suit is unlikely if the defendant could just AI wash the work in the first place.
by davemp
4/10/2026 at 11:28:52 PM
No, a court did not declare that. The case involved a person trying to register a work with only the AI system listed as author. The Supreme Court decided that you can't do that, you need to list a human being as author to register a work with the Copyright Office. This stems from existing precedent where someone tried to register a photograph with the monkey photographer listed as author.I don't believe the idea that humans can or can't claim copyright over AI-authored works has been tested. The Copyright Office says your prompt doesn't count and you need some human-authored element in the final work. We'll have to see.
by tadfisher
4/11/2026 at 1:47:15 AM
It's almost a certainty that you can't copyright code that was generated entirely by an AI.Copyright requires some amount of human originality. You could copyright the prompt, and if you modify the generated code you can claim copyright on your modifications.
The closest applicable case would be the monkey selfie.
https://en.wikipedia.org/wiki/Monkey_selfie_copyright_disput...
by papercrane
4/11/2026 at 3:17:24 AM
It's almost certain that you're wrong. It's like saying I can't copyright a song if my modular synthesizer generated it. Why would you think this?by abletonlive
4/11/2026 at 3:28:08 AM
[dead]by wetpaws
4/11/2026 at 1:36:09 AM
I’m curious to see if subscription vs free ends up mattering here. If it is a work for hire, generally it doesn’t matter how the work was produced, the end result is mine, because I contracted and instructed (prompted?) someone to do it for me. So will the copyright office decide it cares if I paid for the AI tool explicitly?by manwe150
4/11/2026 at 7:08:31 AM
That would depend on whether those who sold you the software-output, had copyright to it.by galaxyLogic
4/10/2026 at 11:23:34 PM
> Didn't a court in the US declare that AI generated content cannot be copyrighted?No, my understanding is that AI generated content can't be copyrighted by the AI. A human can still copyright it, however.
by RussianCow
4/11/2026 at 2:55:03 AM
It's obvious that a computer program cannot have copyright because computer programs are not persons in any currently existing jurisdiction.Whether a person can claim copyright of the output of a computer program is generally understood as depending on whether there was sufficient creative effort from said person, and it doesn't really matter whether the program is Photoshop or ChatGPT.
by Sharlin
4/11/2026 at 3:04:08 AM
Just thinking out loud... why can't an algorithm be an artificial person in the legal sense that a corporation is? Why not legally incorporate the AI as a corporation so it can operate in the real world: have accounts, create and hold copyrights...by paradoxyl
4/11/2026 at 5:41:35 AM
Because the law doesn't say it can. It's that simple.by direwolf20
4/11/2026 at 3:28:57 AM
Corporations are required to have human directors with full operational authority over the corporation's actions. This allows a court to summon them and compel them to do or not do things in the physical world. There's no reason a corporation can't choose to have an AI operate their accounts, but this won't affect the copyright status, and if the directors try to claim they can't override the AI's control of the accounts they'll find themselves in jail for contempt the first time the corporation faces a lawsuit.by SpicyLemonZest
4/11/2026 at 7:11:42 AM
So if creative effort was put into writing the prompt, then whoever wrote the prompt should have the copyright to the output produced by ChatGPT?by galaxyLogic
4/11/2026 at 1:34:06 AM
Public domain code is GPL compatibleby singpolyma3
4/10/2026 at 8:12:52 PM
Same as if a regular person did the same. They are responsible for it. If you're using AI, check the code doesn't violate licensesby afro88
4/10/2026 at 9:27:18 PM
In certain law cases plagiarization can be influenced by the fact if person is exposed to the copyrighted work. AI models are exposed to very large corpus of works..by rzmmm
4/10/2026 at 10:16:50 PM
Copyright infringement and plagiarism are not the same or even very closely related. They're different concepts and not interchangeable. Relative to copyright infringement, cases of plagiarism are rarely a matter for courts to decide or care about at all. Plagiarism is primarily an ethical (and not civil or criminal) matter. Rather than be dealt with by the legal system, it is the subject of codes of ethics within e.g. academia, journalism, etc. which have their own extra-judicial standards and methods of enforcement.by cxr
4/10/2026 at 11:22:34 PM
I suspect they were instead referring to patents; for example, when I worked at Google, they told the engineers not to read patents because then the engineer might invent something infringing, I think it's called willful infringement. No other employer I've worked for has every raised this as an issue, while many lawyers at google would warn against this.by dekhn
4/10/2026 at 8:37:24 PM
As opposed to an irregular person?LLMs are not persons, not even legal ones (which itself is a massive hack causing massive issues such as using corporate finances for political gain).
A human has moral value a text model does not. A human has limitations in both time and memory available, a model of text does not. I don't see why comparisons to humans have any relevance. Just because a human can do something does not mean machines run by corporations should be able to do it en-masse.
The rules of copyright allow humans to do certain things because:
- Learning enriches the human.
- Once a human consumes information, he can't willingly forget it.
- It is impossible to prove how much a human-created intellectual work is based on others.
With LLMs:
- Training (let's not anthropomorphize: lossily-compressing input data by detecting and extracting patterns) enriches only the corporation which owns it.
- It's perfectly possible to create a model based only on content with specific licenses or only public domain.
- It's possible to trace every single output byte to quantifiable influences from every single input byte. It's just not an interesting line of inquiry for the corporations benefiting from the legal gray area.
by martin-t
4/11/2026 at 4:36:03 AM
Dude come on, I clearly wasn't saying LLMs are people. My point was it's a tool and it's the responsibility of the person wielding it to check outputs.If it's too hard to check outputs, don't use the tool.
Your arguments about copyright being different for LLMs: at the moment that's still being defined legally. So for now it's an ethical concern rather than a legal one.
For what it's worth I agree that LLMs being trained on copyright material is an abuse of current human oriented copyright laws. There's no way this will just continue to happen. Megacorps aren't going to lie down if there's a piece of the pie on the table, and then there's precedent for everyone else (class action perhaps)
by afro88
4/10/2026 at 8:21:58 PM
How could you do that though? You can’t guarantee that there aren’t chunks of copied code that infringes.by sarchertech
4/10/2026 at 9:01:36 PM
Let me introduce you to the concept of submarine patents...by Andrex
4/10/2026 at 8:29:24 PM
But the responsible party is still the human who added the code. Not the tool that helped do so.by shevy-java
4/10/2026 at 8:42:17 PM
The practical concern of Linux developers regarding responsibility is not being able to ban the author, it's that the author should take ongoing care for his contribution.by aargh_aargh
4/10/2026 at 8:40:52 PM
That's not going to shield the Linux organization.by Cytobit
4/10/2026 at 10:26:05 PM
A DCO bearing a claim of original authorship (or assertion of other permitted use) isn't going to shield them entirely, but it can mitigate liability and damages.by cxr
4/11/2026 at 1:02:47 AM
Can it though? As far as I know this hasn’t been tested.by sarchertech
4/10/2026 at 8:33:47 PM
In a court case the responsibility party very well could be the Linux foundation because this is a foreseeable consequence of allowing AI contributions. There’s no reasonable way for a human to make such a guarantee while using AI generated code.by sarchertech
4/10/2026 at 8:50:41 PM
It’s not about the mechanism: responsibility is a social construct, it works the way people say that it works. If we all agree that a human can agree to bear the responsibility for AI outputs, and face any consequences resulting from those outputs, then that’s the whole shebang.by Chance-Device
4/10/2026 at 8:59:46 PM
Sure we could change the law. It would be a stupid change to allow individuals, organizations, and companies to completely shield themselves from the consequences of risky behaviors (more than we already do) simply by assigning all liability to a fall guy.by sarchertech
4/10/2026 at 9:09:43 PM
What law exactly are you suggesting needs to be changed? How is this any different from what already happens right now, today?by Chance-Device
4/10/2026 at 9:20:38 PM
Right now it's very easy not to infringe on copyrighted code if you write the code yourself. In the vast majority of cases if you infringed it's because you did something wrong that you could have prevented (in the case where you didn't do anything wrong, inducement creation is an affirmative defense against copyright infringement).That is not the case when using AI generated code. There is no way to use it without the chance of introducing infringing code.
Because of that if you tell a user they can use AI generated code, and they introduce infringing code, that was a foreseeable outcome of your action. In the case where you are the owner of a company, or the head of an organization that benefits from contributors using AI code, your company or organization could be liable.
by sarchertech
4/11/2026 at 7:22:06 AM
So it's a bit as if Linux Organization told its contributors you can bring in infringing code but you must agree you are liable for any infringement?But if a lawsuit was later brought who would be sued? The individual author or the organization? In other words can an organization reduce its liability if it tells its employees "You can break the law as long as you agree you are solely responsible for such illegal actions?
It would seem to me that the employer would be liable if they "encourage" this way of working?
by galaxyLogic
4/10/2026 at 9:35:38 PM
It’s a foreseeable outcome that humans might introduce copyrighted code into the kernel.I think you’re looking for problems that don’t really exist here, you seem committed to an anti AI stance where none is justified.
by Chance-Device
4/10/2026 at 10:02:15 PM
A human has to willingly violate the law for that to happen though. There is no way for a human to use AI generated that doesn't have a chance of producing copyrighted code though. That's just expected.If you don't think this is a problem take a look at the terms of the enterprise agreements from OpenAI and Anthropic. Companies recognize this is an issue and so they were forced to add an indemnification clause, explicitly saying they'll pay for any damages resulting in infringement lawsuits.
by sarchertech
4/11/2026 at 12:29:45 AM
> Right now it's very easy not to infringe on copyrighted code if you write the code yourself.Humans routinely produce code similar to or identical to existing copyrighted code without direct copying.
by johnisgood
4/11/2026 at 12:49:34 AM
They don’t produce enough similar code to infringe frequently. And if they did independent creation is an affirmative defense to copyright infringement that likely doesn’t apply to LLMs since they have the demonstrated capability to produce code directly from their training set.by sarchertech
4/11/2026 at 12:55:32 AM
You have shifted from "very easy not to infringe" to "don't infringe frequently", which concedes the original point that humans can and do produce infringing code without intent.On independent creation: you are conflating the tool with the user. The defense applies to whether the developer had access to the copyrighted work, not whether their tools did. A developer using an LLM did not access the training set directly, they used a synthesis tool. By your logic, any developer who has read GPL code on GitHub should lose independent creation defense because they have "demonstrated capability to produce code directly from" their memory.
LLM memorization/regurgitation is a documented failure mode, not normal operation (nor typical case). Training set contamination happens, but it is rare and considered a bug. Humans also occasionally reproduce code from memory: we do not deny them independent creation defense wholesale because of that capability!
In any case, the legal question is not settled, but the argument that LLM-assisted code categorically cannot qualify for independent creation defense creates a double standard that human-written code does not face.
by johnisgood
4/11/2026 at 5:43:43 AM
And that's not an infringement. Actual copying is the infringement, not having the same code. The most likely way to have the same code is by copying, but it's not the only way.by direwolf20
4/10/2026 at 9:06:55 PM
In this case, the "fall guy" is the person who actually introduced the code in question into the codebase.They wouldn't be some patsy that is around just to take blame, but the actual responsible party for the issue.
by bpt3
4/10/2026 at 9:30:39 PM
Imagine your a factory owner and you need a chemical delivered from across the country, but the chemical is dangerous and if the tanker truck drives faster than 50 miles per hour it has a 0.001% chance per mile of exploding.You hire an independent contractor and tell him that he can drive 60 miles per hour if he wants to but if it explodes he accepts responsibility.
He does and it explodes killing 10 people. If the family of those 10 people has evidence you created the conditions to cause the explosion in order to benefit your company, you're probably going to lose in civil court.
Linus benefits from the increase velocity of people using AI. He doesn't get to put all the liability on the people contributing.
by sarchertech
4/11/2026 at 6:13:15 AM
Cool analogy! Which has nothing to do with the topic in hand.by raincole
4/11/2026 at 12:34:55 AM
That is a nonsensical analogy on multiple levels, and doesn't even support your own argument.by bpt3
4/11/2026 at 12:46:56 AM
Nice rebuttal.by sarchertech
4/11/2026 at 1:01:58 AM
Why would I put much effort into responding to a post like yours, which makes no sense and just shows that you don't understand what you're talking about?by bpt3
4/10/2026 at 10:41:42 PM
Responsibility is an objective fact, not just some arbitrary social convention. What we can agree or disagree about is where it rests, but that's a matter of inference, an inference can be more or less correct. We might assign certain people certain responsibilities before the fact, but that's to charge them with the care of some good, not to blame them for things before they were charged with their care.by lo_zamoyski
4/10/2026 at 10:42:17 PM
Because contributions to Linux are meticulously attributed to, and remain property of, their authors, those authors bear ultimate responsibility. If Fred Foobar sends patches to the kernel that, as it turns out, contain copyrighted code, then provided upstream maintainers did reasonable due diligence the court will go after Fred Foobar for damages, and quite likely demand that the kernel organization no longer distribute copies of the kernel with Fred's code in it.by bitwize
4/11/2026 at 12:53:58 AM
Anyone distributing infringing material can be liable, and it’s unlikely that this technicality will actually would shield anyone.Anyone who thinks they have a strong infringement case isn’t going to stop at the guy who authored the code, they’re going to go after anyone with deep pockets with a good chance of winning.
by sarchertech
4/11/2026 at 10:29:24 AM
> Anyone distributing infringing material can be liableThere is still the "mens rea" principle. If you distribute infringing material unknowingly, it would very likely not result in any penalties.
by Marha01
4/11/2026 at 7:06:01 AM
There is already lots and lots of non-GPL code in the kernel, under dozens of licenses, see https://raw.githubusercontent.com/Open-Source-Compliance/pac...As long as everything is GPLv2-compatible it‘s okay.
by Tomte
4/10/2026 at 9:27:36 PM
Tab complete does not produce copyrightable material either. Yet we don't require software to be written in nano.by noosphr
4/11/2026 at 3:14:09 AM
This is a nice point that I haven't seen before. It's interesting to regress AI to the simplest form and see how we treat it as a test for the more complex cases.by rpdillon
4/10/2026 at 8:13:05 PM
If the output is public domain it's fine as I understand it.by panzi
4/10/2026 at 8:19:37 PM
Makes sense to me. But so anybody can take Public Domain code and place it under GNU Public License (by dropping it into a Linux source-code file) ?Surely the person doing so would be responsible for doing so, but are they doing anything wrong?
by galaxyLogic
4/10/2026 at 8:26:26 PM
> Surely the person doing so would be responsible for doing so, but are they doing anything wrong?You're perfectly at liberty to relicense public domain code if you wish.
The only thing you can't do is enforce the new license against people who obtain the code independently - either from the same source you did, or from a different source that doesn't carry your license.
by robinsonb5
4/10/2026 at 8:38:05 PM
This is correct, and it's not limited to code. I can take the story of Cinderella, create something new out of it, copyright my new work, but Cinderella remains public domain for someone else to do something with.If I use public domain code in a project under a license, the whole work remains under the license, but not the public domain code.
I'm not sure what the hullabaloo is about.
by cwnyth
4/11/2026 at 1:55:47 AM
If someone else uses your exact same prompt to generate the exact same code, can you claim copyright infringement against them? If the output is possible to copyright, then you could claim their prompt is infringement (just like if it reproduced Harry Potter). If it isn’t copyrightable, then the kernel would not have legal standing to enforce the GPL on those lines of code against any future AI reproduction of them. The developers might need to show that the code is licensed under GPL and only GPL, otherwise there is the possibility the same original contributor (eg the AI) did permit the copy. The GPL is an imposed restriction on what the kernel can legally do with any code contributions. That seems legally complicated for some projects—probably not the kernel with the large amount of pre-AI code, but maybe it spells trouble for smaller newer projects if they want to sue over infringement. IANAL.by manwe150
4/11/2026 at 7:54:25 AM
> If someone else uses your exact same prompt to generate the exact same code, can you claim copyright infringement against them?No, because they've independently obtained it from the same source that you did, so their copy is "upstream" of your imposing of a new license.
Realistically, adding a license to public domain work is only really meaningful when you've used it as a starting point for something else, and want to apply your license to the derivative work.
by robinsonb5
4/11/2026 at 5:45:51 AM
Copyright infringement is triggered by the act of copying, not by having the same bytes.by direwolf20
4/10/2026 at 9:05:18 PM
The core thing about licenses, in general, is that they only grant new usage. If you can already use the code because it's public domain, they don't further restrict it. The license, in that case, is irrelevant.Remember that licenses are powered by copyright - granting a license to non-copyrighted code doesn't do anything, because there's no enforcement mechanism.
This is also why copyright reform for software engineering is so important, because code entering the public domain cuts the gordian knot of licensing issues.
by jaggederest
4/10/2026 at 8:36:12 PM
Linux code doesn't have to strictly be GPL-only, it just has to be GPL-compatible.If your license allows others to take the code and redistribute it with extra conditions, your code can be imported into the kernel. AFAIK there are parts of the kernel that are BSD-licensed.
by miki123211
4/10/2026 at 8:32:11 PM
Sqlite’s source code is public domain. Surely if you dropped the sqlite source code into Linux, it wouldn’t suddenly become GPL code? I’m not sure how it worksby sambaumann
4/11/2026 at 5:40:34 AM
The Linux kernel would become a GPLv2-licensed derivative work of SQLite, but that doesn’t matter, because public domain works, by definition, are not subject to copyright restrictions.Claiming copyright on an unmodified public domain work is a lie, so in some circumstances could be an element of fraud, but still wouldn’t be a copyright violation.
by jasomill
4/10/2026 at 8:45:48 PM
This ruling is IMO/IANAL based on lawyers and judges not understanding how LLMs work internally, falling for the marketing campaign calling them "AI" and not understanding the full implications.LLM-creation ("training") involves detecting/compressing patterns of the input. Inference generates statistically probable based on similarities of patterns to those found in the "training" input. Computers don't learn or have ideas, they always operate on representations, it's nothing more than any other mechanical transformation. It should not erase copyright any more than synonym substitution.
by martin-t
4/10/2026 at 9:38:51 PM
>LLM-creation ("training") involves detecting/compressing patterns of the input.There's a pretty compelling argument that this is essentially what we do, and that what we think of as creativity is just copying, transforming, and combining ideas.
LLMs are interesting because that compression forces distilling the world down into its constituent parts and learning about the relationships between ideas. While it's absolutely possible (or even likely for certain prompts) that models can regurgitate text very similar to their inputs, that is not usually what seems to be happening.
They actually appear to be little remix engines that can fit the pieces together to solve the thing you're asking for, and we do have some evidence that the models are able to accomplish things that are not represented in their training sets.
Kirby Ferguson's video on this is pretty great: https://www.youtube.com/watch?v=X9RYuvPCQUA
by supern0va
4/10/2026 at 10:21:17 PM
So? Why should it be legal?If people find this cool and wanna play with it, they can, just make sure to only mix compatible licenses in the training data and license the output appropriately. Well, the attribution issue is still there, so maybe they can restrict themselves to public domain stuff. If LLMs are so capable, it shouldn't limit the quality of their output too much.
Now for the real issue: what do you think the world will look like in 5 or 10 years if LLMs surpass human abilities in all areas revolving around text input and output?
Do you think the people who made it possible, who spent years of their life building and maintaining open source code, will be rewarded? Or will the rich reap most of the benefit while also simultaneously turning us into beggars?
Even if you assume 100% of the people doing intellectual work now will convert to manual work (i.e. there's enough work for everyone) and robots don't advance at all, that'll drive the value of manual labor down a lot. Do you have it games out in your head and believe somehow life will be better for you, let alone for most people? Or have yo not thought about it at all yet?
by martin-t
4/11/2026 at 10:32:11 AM
The best answer to those issues is still Basic Income.by Marha01
4/11/2026 at 7:39:18 AM
> Do you think the people who made it possible, who spent years of their life building and maintaining open source code, will be rewarded?I think they should be rewarded more than they are currently. But isn't the GNU Public License bassically saying you can use such source-code without giving any rewards what so ever?
But I see your The reward for Open Source developers is the public recognition for their works. LLMs can take that recognition away.
by galaxyLogic
4/10/2026 at 9:24:14 PM
fortunately, you aren't only operating on representations, right? lemme check my Schopenhauer right quick...by timmmmmmay
4/10/2026 at 8:28:49 PM
But why should AI then be attributed if it is merely a tool that is used?by shevy-java
4/10/2026 at 10:56:56 PM
Having an honesty based tag could be only way to monitor impact or get after a fix in code bases if things go south.That is at the moment: - Nobody knows for sure what agents might add and their long term effects on codebases.
- It's at best unclear that AI content in a codebase can be reliably determined automatically.
- Even if it's not malicious, at least some of its contributions are likely to be deleterious and pass undetected by human review.
by lonelyasacloud
4/11/2026 at 6:30:05 AM
AI tools can do the entire job from finding the problem, implementing and testing it.It's different from the regular single purpose static tools.
by yrds96
4/11/2026 at 1:47:47 AM
This is a good point but I'd take it in the opposite direction from the implication, we should document which tools were used in general, it'd be a neat indicator of what people use.by hgoel
4/10/2026 at 8:54:48 PM
it makes sense to keep track of what model wrote what code to look for patterns, behaviors, etc.by plmpsu
4/10/2026 at 9:01:51 PM
It isn't?> AI agents MUST NOT add Signed-off-by tags. Only humans can legally certify the Developer Certificate of Origin (DCO).
They mention an Assisted-by tag, but that also contains stuff like "clang-tidy". Surely you're not interpreting that as people "attributing" the work to the linter?
by streetfighter64