this post was submitted on 12 Apr 2026
137 points (92.5% liked)

Technology

83695 readers
5192 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] catlover@sh.itjust.works 9 points 1 hour ago (2 children)

I'd still be highly sceptical about pull requests with code created by llms. Personally what I noticed is that the author of such pr doesn't even read the code, and i have to go through all the slop

[–] terabyterex@lemmy.world 0 points 14 minutes ago

Did we all forget about stackoverflow?

[–] kcuf@lemmy.world 1 points 16 minutes ago

Ya I'm finding myself being the bad code generator at work as I'm scattered across so many things at the moment due to attrition and AI can do a lot of the boilerplate work, but it's such a time and energy sink to fully review what it generates and I've found basic things I missed that others catch and shows the sloppiness. I usually take pride in my code, but I have no attachment to what's generated and that's exposing issues with trying to scale out using this

[–] Blue_Morpho@lemmy.world 77 points 3 hours ago (2 children)

The title of the article is extraordinary wrong that makes it click bait.

There is no "yes to copilot"

It is only a formalization of what Linux said before: All AI is fine but a human is ultimately responsible.

" AI agents cannot use the legally binding "Signed-off-by" tag, requiring instead a new "Assisted-by" tag for transparency"

The only mention of copilot was this:

"developers using Copilot or ChatGPT can't genuinely guarantee the provenance of what they are submitting"

This remains a problem that the new guidelines don't resolve. Because even using AI as a tool and having a human review it still means the code the LLM output could have come from non GPL sources.

[–] marlowe221@lemmy.world 17 points 2 hours ago* (last edited 2 hours ago) (1 children)

Yeah, that’s also my question. Partially because I am a former-lawyer-turned-software-developer… but, yeah. How are the kernel maintainers supposed to evaluate whether a particular PR contains non-GPL code?

Granted, this was potentially an issue before LLMs too, but nowhere near the scale it will be now.

(In the interests of full disclosure, my legal career had nothing to do with IP law or software licensing - I did public interest law).

[–] stsquad@lemmy.ml 5 points 1 hour ago

They don't, just like they don't with human submitted stuff. The point of the Signed-off-by is the author attests they have the rights to submit the code.

[–] anarchiddy@lemmy.dbzer0.com 6 points 2 hours ago

Yup.

I would also just point out that this doesnt change the legal exposure to the Linux kernel to infringing submissions from before the advent of LLMs.

[–] hperrin@lemmy.ca 6 points 1 hour ago (2 children)

This is a bad move. The GPL license cannot be enforced on AI generated code.

[–] terabyterex@lemmy.world 1 points 13 minutes ago

Thats not true. The new article being shoved down lemmy's throat is not correct. They site court cases and come to bad conclusions

[–] Goodlucksil@lemmy.dbzer0.com 1 points 16 minutes ago

AI generated code cannot be copyrighted, can it? Then it can be relicensed as GPL.

[–] theherk@lemmy.world 78 points 3 hours ago (2 children)

Seems like a reasonable approach. Make people be accountable for the code they submit, no matter the tools used.

[–] hperrin@lemmy.ca 4 points 1 hour ago (1 children)

No, it’s not a reasonable approach. Make people be the authors of the code they submit is reasonable, because then it can be released under the GPL. AI generated code is public domain.

[–] theherk@lemmy.world 2 points 40 minutes ago

I suppose there should be no code generators, assemblers, compilers, linkers, or lsp’s then either? Just etching 1’s and 0’s?

[–] ell1e@leminal.space 11 points 2 hours ago (1 children)

If the accountability cannot be practically fulfilled, the reasonable policy becomes a ban.

What good is it to say "oh yeah you can submit LLM code, if you agree to be sued for it later instead of us"? I'm not a lawyer and this isn't legal advice, but sometimes I feel like that's what the Linux Foundation policy says.

[–] ViatorOmnium@piefed.social 12 points 2 hours ago (4 children)

But this was already the case. When someone submitted code to Linux they always had to assume responsibility for the legality of the submitted code, that's one of the points of mandatory Signed-off-by.

load more comments (4 replies)
[–] 0ndead@infosec.pub 26 points 3 hours ago (3 children)

“Yes to Copilot, no to AI slop”

Pick One

[–] truthfultemporarily@feddit.org 10 points 3 hours ago (10 children)

Where does slop start? If you use auto complete and it is just adding a semicolon or some braces, is it slop? Is producing character by character what you would have wrote yourself slop?

How about using it for debugging?

[–] badgermurphy@lemmy.world 1 points 1 hour ago

There's the rub. When establishing laws and guidelines, every term must be explicitly defined. Lack of specificity in these definitions is where bad-faith actors hide their misdeeds by technically obeying the letter of the law due to its vagueness, while flagrantly violating its spirit.

Its why today, in the USA, corporations are legally people when its convenient, and not when its not, and the expenditure of money is governments protected "free speech".

[–] hperrin@lemmy.ca 2 points 1 hour ago

You don’t need AI to autocomplete code. We’ve had autocomplete for over 30 years.

[–] ell1e@leminal.space 4 points 2 hours ago* (last edited 2 hours ago)

If you would have written it yourself the same way, why not write it yourself? (And there was autocomplete before the age of LLMs, anyway.)

The big problems start with situations where it doesn't match what you would have written, but rather what somebody else has written, character by character.

load more comments (7 replies)
[–] femtek@lemmy.blahaj.zone 8 points 3 hours ago

I mean I don't use copilot but a self hosted Claude at work for debugging and creating templates. I still run thru and test it. I'm only doing crossplane, kyverno, kubernetes infra things though and I started without it so I have an understanding. Now running their someone's crossplane composition written in go and I asked them about this error and he just said get the AI to fix it was worrying since his last day is next week.

load more comments (1 replies)
[–] ell1e@leminal.space 17 points 3 hours ago* (last edited 3 hours ago) (6 children)

Ultimately, the policy legally anchors every single line of AI-generated code

How would that even be possible? Given the state of things:

https://dl.acm.org/doi/10.1145/3543507.3583199

Our results suggest that [...] three types of plagiarism widely exist in LMs beyond memorization, [...] Given that a majority of LMs’ training data is scraped from the Web without informing content owners, their reiteration of words, phrases, and even core ideas from training sets into generated texts has ethical implications. Their patterns are likely to exacerbate as both the size of LMs and their training data increase, [...] Plagiarized content can also contain individuals’ personal and sensitive information.

https://www.theatlantic.com/technology/2026/01/ai-memorization-research/685552/

Four popular large language models—OpenAI’s GPT, Anthropic’s Claude, Google’s Gemini, and xAI’s Grok—have stored large portions of some of the books they’ve been trained on, and can reproduce long excerpts from those books. [...] This phenomenon has been called “memorization,” and AI companies have long denied that it happens on a large scale. [...]The Stanford study proves that there are such copies in AI models, and it is just the latest of several studies to do so.

https://www.twobirds.com/en/insights/2025/landmark-ruling-of-the-munich-regional-court-(gema-v-openai)-on-copyright-and-ai-training

The court confirmed that training large language models will generally fall within the scope of application of the text and data mining barriers, [...] the court found that the reproduction of the disputed song lyrics in the models does not constitute text and data mining, as text and data mining aims at the evaluation of information such as abstract syntactic regulations, common terms and semantic relationships, whereas the memorisation of the song lyrics at issue exceeds such an evaluation and is therefore not mere text and data mining

https://www.sciencedirect.com/science/article/pii/S2949719123000213#b7

In this work we explored the relationship between discourse quality and memorization for LLMs. We found that the models that consistently output the highest-quality text are also the ones that have the highest memorization rate.

https://arxiv.org/abs/2601.02671

recent work shows that substantial amounts of copyrighted text can be extracted from open-weight models. However, it remains an open question if similar extraction is feasible for production LLMs, given the safety measures [...]. We investigate this question [...] our work highlights that, even with model- and system-level safeguards, extraction of (in-copyright) training data remains a risk for production LLMs.

How does merely tagging the apparently stolen content make it less problematic, given I'm guessing it still won't have any attribution of the actual source (which for all we know, might often even be GPL incompatible)?

But I'm not a lawyer, so I guess what do I know. But even from a non-legal angle, what is this road the Linux Foundation seems to embrace of just ignoring the license of projects? Why even have the kernel be GPL then, rather than CC0?

I don't get it. And the article calling this "pragmatism" seems absurd to me.

load more comments (6 replies)
[–] veniasilente@lemmy.dbzer0.com 3 points 2 hours ago

How is this all supposed to be, when AI code can not be copyrighted and thus those submissions to the Linux kernel can not be eg.: GPLv{number}?

[–] mesamunefire@piefed.social 5 points 2 hours ago

I hate ai in my kernel....

The rule should be "if you get caught using LLMs or caling them 'AI', you're a dipshit and will never ever be let near the Kernel again."

[–] twinnie@feddit.uk 11 points 3 hours ago (8 children)

No point getting upset about this, it’s inevitable. So many FOSS programmers work thanklessly for hours and now there’s some tool to take loads of that work away, of course they’re going to use it. I know loads of people complain about it but used responsibly it can take care of so much of the mundane work. I used to spend 10% of my time writing code then 90% debugging it. If I do that 10% then give it to Claude to go over I find it just works.

[–] uuj8za@piefed.social 1 points 1 hour ago

but used responsibly

That's like the most incredibly hard part of all of this. Everything is aligned so that you don't use it responsibly. And it's really hard to guard against this.

Just a few days ago, I was pairing with a coworker and he was using Claude to do a bunch of stuff. He didn't check any of it. I thought he was gonna check stuff before pushing stuff... And nope! I said, "Wait, shouldn't we review the changes to make sure they're correct?" And he said, "Nah, it's probably fine. I trust it. Plus, even if it's wrong, we'll just blame the AI and we can just fix it later."

...

Yes, checking the work would have negated all of the "time saved" and he was being a lazy fuck.

People who don't like coding or engineering use this and they are not interested in using this responsibly.

load more comments (7 replies)
[–] robinadams@lemmy.wtf 2 points 1 hour ago

Well, time to switch to NetBSD

[–] XLE@piefed.social 10 points 3 hours ago (3 children)

This seems like an ill-thought-out decision, especially in a landscape where Linux should be differentiating itself from, and not following Windows.

The titular "slop" just means "bad AI generated code is banned" but the definition of "bad" is as vague as Google's "don't be evil." Good luck enforcing it, especially in an open-source project where people's incentives aren't tied to a paycheck.

Title is also inaccurate regarding CoPilot (the Microsoft brand AI tool), as a comment there mentions

says yes to Copilot

Where in the article does it say that?? The only mention of CoPilot is where it talks about LLM-generated code having unverifiable provenance. Reply

[–] Naich@piefed.world 10 points 3 hours ago

Google's "don't be evil" was like a warrant canary. It didn't need to be precise, it just needed to be there.

[–] avidamoeba@lemmy.ca 6 points 3 hours ago (2 children)

They're already enforcing it. PRs are reviewed and bad ones are rejected all the time.

load more comments (2 replies)
load more comments (1 replies)
[–] treadful@lemmy.zip 3 points 2 hours ago (2 children)

I'm curious how this is going to play out legally for copyright. If you accept AI code, you can't copyright it, so aren't you essentially forfeiting the copyleft license?

load more comments (2 replies)
load more comments
view more: next ›