this post was submitted on 12 Apr 2026
574 points (94.8% liked)

Technology

83753 readers
2862 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] theherk@lemmy.world 139 points 1 day ago (2 children)

Seems like a reasonable approach. Make people be accountable for the code they submit, no matter the tools used.

[–] ell1e@leminal.space 26 points 1 day ago (1 children)

If the accountability cannot be practically fulfilled, the reasonable policy becomes a ban.

What good is it to say "oh yeah you can submit LLM code, if you agree to be sued for it later instead of us"? I'm not a lawyer and this isn't legal advice, but sometimes I feel like that's what the Linux Foundation policy says.

[–] ViatorOmnium@piefed.social 49 points 1 day ago (1 children)

But this was already the case. When someone submitted code to Linux they always had to assume responsibility for the legality of the submitted code, that's one of the points of mandatory Signed-off-by.

[–] hperrin@lemmy.ca 6 points 1 day ago (2 children)

No, it’s not a reasonable approach. Make people be the authors of the code they submit is reasonable, because then it can be released under the GPL. AI generated code is public domain.

[–] theherk@lemmy.world 3 points 1 day ago (1 children)

I suppose there should be no code generators, assemblers, compilers, linkers, or lsp’s then either? Just etching 1’s and 0’s?

[–] hperrin@lemmy.ca 5 points 21 hours ago (1 children)

The copyright office has made it explicitly clear that those tools do not interfere with the traditional elements of authorship, and that the use of LLMs does. So, if you don’t want to take my word for it, take the US Copyright Office’s word for it.

[–] theherk@lemmy.world 1 points 18 hours ago* (last edited 17 hours ago) (1 children)

As the agency overseeing the copyright registration system, the Office has extensive experience in evaluating works submitted for registration that contain human authorship combined with uncopyrightable material, including material generated by or with the assistance of technology. It begins by asking “whether the ‘work’ is basically one of human authorship, with the computer [or other device] merely being an assisting instrument, or whether the traditional elements of authorship in the work (literary, artistic, or musical expression or elements of selection, arrangement, etc.) were actually conceived and executed not by man but by a machine.” In the case of works containing AI-generated material, the Office will consider whether the AI contributions are the result of “mechanical reproduction” or instead of an author’s “own original mental conception, to which [the author] gave visible form.” The answer will depend on the circumstances, particularly how the AI tool operates and how it was used to create the final work. This is necessarily a case-by-case inquiry. If a work’s traditional elements of authorship were produced by a machine, the work lacks human authorship and the Office will not register it For example, when an AI technology receives solely a prompt from a human and produces complex written, visual, or musical works in response, the “traditional elements of authorship” are determined and executed by the technology—not the human user. Based on the Office’s understanding of the generative AI technologies currently available, users do not exercise ultimate creative control over how such systems interpret prompts and generate material. Instead, these prompts function more like instructions to a commissioned artist—they identify what the prompter wishes to have depicted, but the machine determines how those instructions are implemented in its output. For example, if a user instructs a text-generating technology to “write a poem about copyright law in the style of William Shakespeare,” she can expect the system to generate text that is recognizable as a poem, mentions copyright, and resembles Shakespeare’s style. But the technology will decide the rhyming pattern, the words in each line, and the structure of the text. When an AI technology determines the expressive elements of its output, the generated material is not the product of human authorship. As a result, that material is not protected by copyright and must be disclaimed in a registration application.

In other cases, however, a work containing AI-generated material will also contain sufficient human authorship to support a copyright claim. For example, a human may select or arrange AI-generated material in a sufficiently creative way that “the resulting work as a whole constitutes an original work of authorship.” Or an artist may modify material originally generated by AI technology to such a degree that the modifications meet the standard for copyright protection. In these cases, copyright will only protect the human-authored aspects of the work, which are “independent of ” and do “not affect” the copyright status of the AI-generated material itself.

This policy does not mean that technological tools cannot be part of the creative process. Authors have long used such tools to create their works or to recast, transform, or adapt their expressive authorship. For example, a visual artist who uses Adobe Photoshop to edit an image remains the author of the modified image, and a musical artist may use effects such as guitar pedals when creating a sound recording. In each case, what matters is the extent to which the human had creative control over the work’s expression and “actually formed” the traditional elements of authorship.

https://www.copyright.gov/ai/ai_policy_guidance.pdf

What this makes clear is that it certainly isn’t black or white as you say. Nevertheless, automation converting an input to an output, simply cannot be the only mechanism used in determining authorship.

And that wouldn’t change my statement anyway, but rather supports it. The person submitting a patch must be accountable for its contents.

An outright ban would need to carefully define how an input gets converted to an output, and that may not be so clear. To be effectively clear, one would have to potentially end the use of many tools that have been used for many years in the kernel, including snippet generation, spelling and grammar correction, IDE autocompleting. So such a reductive view simply will not suffice.


Additionally, copywritability and licenseability are wholly different questions. And it does not violate GPL to include public domain content, since the license applies to the aggregate work.

[–] hperrin@lemmy.ca 3 points 17 hours ago* (last edited 17 hours ago) (1 children)

If a work’s traditional elements of authorship were produced by a machine, the work lacks human authorship and the Office will not register it For example, when an AI technology receives solely a prompt from a human and produces complex written, visual, or musical works in response, the “traditional elements of authorship” are determined and executed by the technology—not the human user. Based on the Office’s understanding of the generative AI technologies currently available, users do not exercise ultimate creative control over how such systems interpret prompts and generate material. Instead, these prompts function more like instructions to a commissioned artist—they identify what the prompter wishes to have depicted, but the machine determines how those instructions are implemented in its output. For example, if a user instructs a text-generating technology to “write a poem about copyright law in the style of William Shakespeare,” she can expect the system to generate text that is recognizable as a poem, mentions copyright, and resembles Shakespeare’s style. But the technology will decide the rhyming pattern, the words in each line, and the structure of the text. When an AI technology determines the expressive elements of its output, the generated material is not the product of human authorship. As a result, that material is not protected by copyright and must be disclaimed in a registration application.

That seems very clear to me. Generative AI output is not human authored, and therefore not copyrighted.

The policy I use also makes very clear the definition of AI generated material:

https://sciactive.com/human-contribution-policy/#Definitions

I’m not exactly sure how you can possibly think there is an equivalence between a tool like a spelling and grammar checker and a generative AI, but there’s a reason the copyright office will register works that have been authored using spelling and grammar checkers, but not works that have been authored using LLMs.

[–] theherk@lemmy.world 2 points 17 hours ago (1 children)

Just read the next two paragraphs. Don’t just stop because you got to something that you like. The equivalence I draw is clear. You don’t like it, and that’s okay. But one would have to clarify exactly what the ban entails, and that wouldn’t be as clear as you might think. LLM’s only, transformers specifically, what about graph generation, other ML models? Is it just ML? If so, is that because a matrix lattice was used to get from input to output? Could other deterministic math functions trigger the same ban? What is a spell checker used RNG to select best replacement from a list of correct options? What if a compiler introduces an assembled output with an optimization not of the authors writing?

Do you see why they say “The answer will depend on the circumstances, particularly how the AI tool operates and how it was used to create the final work. This is necessarily a case-by-case inquiry”?

And that still affects copywriteability, not license compliance.

[–] hperrin@lemmy.ca 1 points 17 hours ago* (last edited 16 hours ago) (1 children)

Do you want to explain to me what, in those two paragraphs, means that the use of spell checkers and LLMs is equivalent with regard to copyrightability? It seems like those paragraphs make it clear that the use of spell checkers is not the same as LLMs.

The policy I use bans “generative AI model” output. Generative AI is a pretty well defined term:

https://en.wikipedia.org/wiki/Generative_AI

https://www.merriam-webster.com/dictionary/generative%20AI

If you have trouble determining whether something is a generative AI model, you can usually just look up how it is described in the promotional materials or on Wikipedia.

Type: Large language model, Generative pre-trained transformer

- https://en.wikipedia.org/wiki/Claude_(language_model)

I never said it violates GPL to include public domain code. I’m not sure where you got that from. What I said is that public domain code can’t really be released under the GPL. You can try, but it’s not enforceable. As in, you can release it under that license, but I can still do whatever I want with it, license be damned, because it’s public domain.

I did that with this vibe coded project:

https://github.com/hperrin/gnata

I just took it and rereleased it as pubic domain, because that’s what it is anyway.

[–] eah@programming.dev 2 points 4 hours ago (1 children)

public domain code can’t really be released under the GPL

Disney created films based on old fairy tales. Disney has a copyright on those films even though they include elements from the public domain because the films also include the artists' original expression. The linux kernel (probably) contains public domain AI-generated code alongside original work from its many contributors. If you wanted to get the entire project into the public domain, you'd have to get permission from nearly all its contributors or wait for their copyright term to expire. The small snippets of code which were AI-generated are public domain. The bulk of the project isn't, and the project as a whole isn't.

As much as I dislike AI, I can't say I understand forbidding AI-generated contributions on the grounds that the submitted code is public domain. I suppose somebody can come along and "steal" the public domain snippets, but I suspect it's difficult to definitively tell apart the human-written code from AI-generated and strip out the human-written bits. If they do, what's the issue? It wasn't yours to begin with and you can still keep it in your project. Moreover, now that the magical plagiarism machines exist, who's going to be lifting code in this way, anyway?

[–] hperrin@lemmy.ca 1 points 3 hours ago

I mean, yeah, you can make the argument that owning the copyrights to all of the code in your project isn’t important. I don’t agree, but that’s certainly a valid stance. Apparently the Linux maintainers are on your side. That makes me sad. Copyright ownership of the things I produce is very important to me.

[–] ziproot@lemmy.ml 1 points 1 day ago (1 children)

Isn’t that the rule? The author has to be a human?

The new guidelines mandate that AI agents cannot use the legally binding "Signed-off-by" tag, requiring instead a new "Assisted-by" tag for transparency. Ultimately, the policy legally anchors every single line of AI-generated code and any resulting bugs or security flaws firmly onto the shoulders of the human submitting it.

[–] hperrin@lemmy.ca 2 points 20 hours ago

If the author is an LLM, then the author is not a human.