this post was submitted on 14 Jan 2024
837 points (99.2% liked)

Technology

59605 readers
3415 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] General_Effort@lemmy.world 5 points 10 months ago (2 children)
[–] TWeaK@lemm.ee 9 points 10 months ago (2 children)

So where does that say I'm wrong?

I said fair use covers news, education, research, criticism, or comment.

for purposes such as criticism, comment, news reporting, teaching (including multiple copies for classroom use), scholarship, or research

Then I said the next thing considered is whether it is commercial.

In determining whether the use made of a work in any particular case is a fair use the factors to be considered shall include— (1) the purpose and character of the use, including whether such use is of a commercial nature or is for nonprofit educational purposes

I didn't cover everything in the law, I just covered the relevant points in a way that could be easily understood and related to the subject at hand.

My point is that the copying AI does isn't really research, but even if it were considered research it is absolutely commercial and thus should not have a fair use exemption.

[–] General_Effort@lemmy.world 3 points 10 months ago* (last edited 10 months ago) (2 children)

You need to read this carefully. It's a statute. It means exactly what it says.

purposes such as criticism, comment, news reporting, teaching (including multiple copies for classroom use), scholarship, or research

Such as means that these are examples. This is not a complete list.

the factors to be considered shall include

All of these factors must be considered. It does not mean that other factors cannot be considered. These are not categories.

A commercial purpose does not rule out a finding of fair use (and vice versa). It must be considered and that is all.

I don't think that Meta's use can be classed as commercial. Presumably, they do hope that the research budget will pay off eventually. But what must be considered is the particular copying in question. Llama 2's license looks to me fairly non-commercial.


Eventually, fair use derives from the constitution. Copyright is a limitation on the freedom of the press (and of speech). But it cannot completely do away with these freedoms. The examples given in the statue here could not be banned completely even if they were not mentioned.

The US Constitution itself allows congress to create copyrights. Or more precisely, it empowers congress to promote the Progress of Science and useful Arts by creating copyrights. That's another limitation.

I've seen a number of far-right commenters admit that this money grab would harm AI development (a "useful Art"). I think mostly these commenters hold some far-right ideology à la Ayn Rand that values property over society, but some may just be selfish and believe that they would personally benefit. Either way, it's straight up anti-constitutional.

[–] wikibot@lemmy.world 2 points 10 months ago

Here's the summary for the wikipedia article you mentioned in your comment:

The Copyright Clause (also known as the Intellectual Property Clause, Copyright and Patent Clause, or the Progress Clause) describes an enumerated power listed in the United States Constitution (Article I, Section 8, Clause 8). The clause, which is the basis of copyright and patent laws in the United States, states that: [the United States Congress shall have power] To promote the Progress of Science and useful Arts, by securing for limited Times to Authors and Inventors the exclusive Right to their respective Writings and Discoveries.

^to^ ^opt^ ^out^^,^ ^pm^ ^me^ ^'optout'.^ ^article^ ^|^ ^about^

[–] TWeaK@lemm.ee -1 points 10 months ago (1 children)

Such as means that these are examples. This is not a complete list.

AI developers have explicitly envoked the research exemption. That is why I focused on that. I disagree that what they do is "research" for the reasons I gave previously. Bringing up the fact there are other exemptions is beside the point - they aren't claiming any other exemption!

All of these factors must be considered. It does not mean that other factors cannot be considered. These are not categories.

Sure, but I never said that commerciality was the only thing that should be considered. My claim here is simply that it is so overwhelmingly commercial in nature that it overrides anything else and thus they should not be awarded the privilege of an exemption.

A commercial purpose does not rule out a finding of fair use (and vice versa).

A commercial purpose might not rule out of a finding of fair use. That does not mean it cannot rule out such a finding. All factors must be considered, but any one factor can outweigh the others.

I never said it was an exclusive category, I just brought it up as the most significant factor - one which is not reasonably overruled by any of the others in this circumstance. In fact, every one of those arguably fails. To give detail:

  1. The copying is done in a commercial nature. They sell AI services. It's offered very cheap right now - even for free for limited personal use - but eventually that will change as their demand for profit grows.
  2. The nature of the copied work is varied and includes all kinds of work, commercial and non-commercial. The copying is pandemic.
  3. The whole work has been copied into the training database. Significant portions of the work can and have been reproduced by the finished product, in spite of the finished product allegedly not containing the original work in its database. Furthermore, even if a human genuinely believes they aren't copying something they read before, that does not mean they are innocent of copyright infringement - it is the similarity of the two works that make the determining factor.
  4. AI work is already flooding the market and pushing out original creators. Childrens' books is one area where this is happening extensively - not only does this make it harder for genuine authors to get a break in the market, but they're effectively training children to think AI work is normal. It's not hard to see us headed to a future where people think AI is "real" and original work is "fake", simply by volume.

I will admit, not all of those arguments are very strong (particularly 4.). However 1. is the strongest and I think overrides any argument the other way for any other.

I don’t think that Meta’s use can be classed as commercial. Presumably, they do hope that the research budget will pay off eventually.

Those two statements contradict one another. Of course they want it to be commercial eventually - or, rather, they want to eventually turn a profit. Hell, AI is already being used in a commercial manner: if you want to make significant or non-personal use of AI systems currently on the market, you have to pay for it.

Eventually, fair use derives from the constitution.

Setting aside the fact that AI extends far beyond the borders of the US and its constitution, fair use and copyright are derived from copyright law, which is written by Congress. The Constitution grants Congress the right to write such laws, but no one is "invoking the Constitution" when they enforce copyright or claim fair use. The Constitution gives permission, but the law forms the definition.

AI is not simply a "useful Art". It is a commercial venture that exploits original work without duly compensating the authors of said work. Congress has a greater duty to protect those original authors than it does a business that seeks to exploit their work. I say this as someone who has never really made much of anything original myself. I play a bit of music, but don't compose and just do covers. I probably (lol limewire definitely) infringe on copyright - but I do so exclusively in a non-commercial manner.

Blurting out "far-right" is borderline a personal insult - one which is laughably far from the mark when addressed towards me - and points to you clutching at straws to cling to a frivilous argument.


I now feel the need to ask, why do you so passionately defend AI businesses here? Why do you support them?

Are you that infatuated with the novelty of their product that you have let go of objectivity?


I also have to emphasise again that I'm a little disgusted that you made this political. You've tried to build an argument that "it is a Constitutional right" to infringe copyright in order to have AI tools, and you're implying that anyone who opposes that idea is some kind of far-right nutjob. I hadn't even heard of Ayn Rand before you mentioned her, but have you actually read her work, or did you just watch the Atlas Shrugged movie and form your opinions from internet memes?

I'd actually probably agree with you about AI - if it was non-commercial in nature and truly for the benefit of the people. As it is, I think you are blinded by the sheen of a new toy, without realising it's coated in lead paint.

[–] TWeaK@lemm.ee 1 points 10 months ago

A commercial purpose might not rule out of a finding of fair use.

ARRRRG I spent so long reviewing this comment, over and over and over again, and still there were words wrong. I'm not editing it though, I want the comment to stay clean.

[–] archomrade@midwest.social 2 points 10 months ago* (last edited 10 months ago) (2 children)

Pretty sure @General_Effort@lemmy.world is referring to this portion:

(3) the amount and substantiality of the portion used in relation to the copyrighted work as a whole

(4) the effect of the use upon the potential market for or value of the copyrighted work.

The main argument for this being fair use is both that a single work of copyright bears little to no relationship to the end product, and that the model itself does not effect the market for - or value of - the copyrighted work (note: the market for additional works produced is not what is in question here, it is the market for the work that has been copied).

[–] TWeaK@lemm.ee 1 points 10 months ago

Sorry for the double reply, but I did also expand further upon (3) and (4), and other aspects, in my latest reply to /u/General_Effort@lemmy.world (link to your instance's version): https://midwest.social/comment/6225045

[–] TWeaK@lemm.ee 0 points 10 months ago (1 children)

The main argument for this being fair use is both that a single work of copyright bears little to no relationship to the end product

It bears relationship to the end product when the end product reproduces the original work.

that the model itself does not effect the market for - or value of - the copyrighted work

Given that AI is poised to take over the position of original writers and flood the market with fake work, copying not only their words but their very style, I'd argue it does affect the value of existing work. With children's books already being heavily written by AI, it seems quite likely that we will before too long get to the point where people expect things to be written by AI, thus devaluing true creative and original work.

[–] archomrade@midwest.social 1 points 10 months ago (1 children)

I appreciate your enthusiasm here, but the law (and precedent reading of the law) simply does not bear out a clear interpretation like you're suggesting.

It bears relationship to the end product when the end product reproduces the original work.

This is not how copyright has been applied when speaking of other machine learning processes using logical regression that is considered fair use, as in Text and Data Mining classifications(TDM) (proposed class 7(a) and 7(b) (page 102) in Recommendation of the Register of Copyrights 2021). The model itself is simply a very large regression model, that has created metadata analysis from unstructured data sources. When determining weather an LLM fits into this fair use category, they will look at what the model is and how it is created, not to whether it can be prompted to recreate a similar work. To quote from Comments in Response of Notice of Inquiry on the matter:

Understanding the process of training foundation models is relevant to the generative AI systems’ fair use defenses because the scope of copyright protection does not extend to “statistical information” such as “word frequencies, syntactic patterns, and thematic markers.” Processing in-copyright works to extract “information about the original [work]” does not infringe because it does not “replicat[e] protected expression.

Granted, what is novel about this particular case (LLM's generally) is their apparent ability to re-construct substantially similar works from the same overall process of TDM. Acknowledged, but to borrow again from the same comments as above:

Yet, in limited situations, Generative AI models do copy the training data.24 So unlike prior copy-reliant technologies that courts have held are fair use, it is impossible to say categorically that inputs and outputs of Generative AI will always be fair use. We note in addition that some have argued that the ability of Generative AI to produce artifacts that could pass for human expression and the potential scale of such production may have implications not seen in previous non-expressive use cases. The difficulty with such arguments is that the harm asserted does not flow from the communication of protected expression to any human audience.

Basically, they are asserting that applying copyright to this use that falls outside of its explicit scope would not prevent the same harm caused by that same technology created without the use of the copyrighted works. Any work sufficiently described in publicly available text data could be reconstructed by a sufficiently weighted regression model and the correct prompting. E.g. - if I described a desired output sufficiently enough in my input to the model, the output could be substantially similar to a protected work, regardless of its lack of representation in the training data.

I happen to agree that these AI models represent a threat to the work and livelihoods of real artists, and that the benefit as currently captured by billion-dollar companies is a substantial problem that must be addressed, but I simply do not think the application of copyright in this manner is appropriate (as it will prevent legitimate uses of the technology), nor do i think it is sufficiently preventative in future consolidation of wealth by the use of these models.

Nevermind my personal objections to copyright law on the basis of my worldview - I just don't think copyright is the correct tool to use for the desired protection.

[–] TWeaK@lemm.ee 1 points 10 months ago (1 children)

This is not how copyright has been applied when speaking of other machine learning processes using logical regression that is considered fair use, as in Text and Data Mining classifications(TDM) (proposed class 7(a) and 7(b) (page 102) in Recommendation of the Register of Copyrights 2021).

Your link is merely proposed recommendations. That is not legislation nor case law. Also, the sections on TDM that you reference clearly state (my emphasis):

for the purpose of scholarly research and teaching.

I think this is even more abundantly clear that the research exemption does not apply. AI "research" is in no way "scholarly", it is commercial product development and thus does not align with fair use copyright exemptions.

It's also not talking about building AI, but circumventing DRM in order to preserve art. They're saying that there should be an exemption to the illegal practice of circumventing DRM in certain, limited circumstances. However, they're still only suggesting this! So not only does this not apply to your argument, it isn't even actually in force.

To put your other link into context, this also is not law, but comments from legal professors.

Understanding the process of training foundation models is relevant to the generative AI systems’ fair use defenses because the scope of copyright protection does not extend to “statistical information” such as “word frequencies, syntactic patterns, and thematic markers.” Processing in-copyright works to extract “information about the original [work]” does not infringe because it does not “replicat[e] protected expression.

The flaw here is that the work isn't processed in situ, it is copied into a training database, then processed. The processing may be fine, but the copying is illegal.

If they had a legitimate license to view the work for their purpose, and processed in situ, that might be different.

The difficulty with such arguments is that the harm asserted does not flow from the communication of protected expression to any human audience.

The argument here is that, while it sometimes infringes copyright, the harm it causes isn't primarily from the infringing act. Not always, though that depends. If AI is used to pass off as someone else, then the AI manufacturer has built a tool that facilitates an illegal act, by copying the original work.

However, this, again, ignores the fact that the commercial enterprise has copied the data into their training database without duly compensating the rightsholder.

[–] archomrade@midwest.social 2 points 10 months ago (1 children)

Your link is merely proposed recommendations. That is not legislation nor case law.

It’s also not talking about building AI, but circumventing DRM in order to preserve art. They’re saying that there should be an exemption to the illegal practice of circumventing DRM in certain, limited circumstances. However, they’re still only suggesting this! So not only does this not apply to your argument, it isn’t even actually in force.

At the bottom of the document, the Library of Congress approves all recommendations and adopts them as legal defenses against copyright claims. This is established law, not merely recommendations. Please understand the legal processes we're discussing here.

Regardless, I'm not arguing that this exemption class 7(a) and 7(b) actually apply to AI and LLM's, only that they serve as precedent guidance on how they should be treated in any suit raised. Granted, OpenAI is not a research institution, so this classification would not apply on those grounds, but the way they treat the work being challenged is still relevant. LLM's are transformative in nature. Their use and nature are distinctly similar to that of a searchable database described in Authors Guild, Inc. v. HathiTrust and Authors Guild v. Google (the legal strength is even greater here, since LLM outputs are creative, and do not provide 'copied' expressions as a matter of course - fringe cases not withstanding), and as such we have no reason to expect they'd view it differently in the case of an LLM. Training data is a utilitarian precursor to an expressive tool, as repeatedly affirmed as fair use in existing precedent.

The flaw here is that the work isn’t processed in situ, it is copied into a training database, then processed. The processing may be fine, but the copying is illegal

Fair use describes exemptions to the illegality of unauthorized copies, it is explicitly asserting the copying as legal for a given use. See Authors Guild, Inc. v. HathiTrust and Authors Guild v. Google for reference. Worthy to point out the distinction between a right to control unauthorized use and unauthorized access, and admittedly this would be the weakest point in Meta's case. However, I share the paper author's perspective on illicit sources:

On the other hand, as Michael Carroll argues, there are strong arguments to be made that copying from an infringing source may still be fair use. Carroll argues that '[t]reating an otherwise fair use as unfair because it was made from an infringing source would lead a court to deny the public access to the products of secondary uses that fair use is designed to encourage.' He notes that significant doubt exists as to whether good faith is a consideration in fair use at all. Judge Pierre Leval has also persuasively argued that using a good faith inquiry in fair use analysis 'produces anomalies that conflict with the goals of copyright and adds to the confusion surrounding the doctrine.' Moreover, even if good faith is part of the broader fair use calculus, courts have found that knowing use of an infringing source is not bad faith when the user acts in the reasonable belief that their use is a fair use. There is no recognized 'fruit of the poisonous tree' doctrine in copyright law.

The argument being proposed in the paper (for once, you are correct that this is not established law) is that in other, different cases where TDM is used as a precursor to expressive use, the collection of data for that purpose has been found to be lawful (provided sufficient security is used to prevent infringing, non-exempt abuses). However, the issue we're discussing is novel. The paper is proposing frameworks for how to apply existing precedent to the novel use-case being investigated. There is no case-law to refer to that addresses this specific situation. I can't tell if you're just trying to debate-bro me or actually discuss the merits of the case, but i'd just remind you that none of this is settled, nor am I suggesting it is. My perspective is that precedent supports training data for LLM's as a fair use, and that strengthening copyright in the way proposed does not mitigate the harm being claimed by plaintiffs, and in fact increases harm to the greater public by gatekeeping access to automation tools and consolidating the benefits to already gigantic companies.

If AI is used to pass off as someone else, then the AI manufacturer has built a tool that facilitates an illegal act, by copying the original work.

That's not an issue for copyright, but I agree it ought to be addressed. Once again, the harm doesn't stem from the use of copyrighted material, it stems from the technology itself (the harm doesn't change weather the material is authorized or not, nor does it change to whom harm is done). I really have to stress again that the issues and concerns being raised over AI cannot be sufficiently addressed through the use of copyright law.

[–] TWeaK@lemm.ee 1 points 10 months ago

At the bottom of the document, the Library of Congress approves all recommendations and adopts them as legal defenses against copyright claims. This is established law, not merely recommendations.

Thank you for the clarification.

Their use and nature are distinctly similar to that of a searchable database described in Authors Guild, Inc. v. HathiTrust and Authors Guild v. Google (the legal strength is even greater here, since LLM outputs are creative, and do not provide ‘copied’ expressions as a matter of course - fringe cases not withstanding), and as such we have no reason to expect they’d view it differently in the case of an LLM. Training data is a utilitarian precursor to an expressive tool, as repeatedly affirmed as fair use in existing precedent.

This is indeed a complicated subject, and thank you again for your insight. These are very good example cases, because Google's searchable book database is exactly the same as the training databases LLM's use to develop their transform nodes.

The difference between the Authors Guild cases and this one, as I see it, is that Google and HathiTrust are acting to preserve information and art for future generations - there is an inherent benefit to society front and centre with their goals. With LLM's, the goal is to develop a commercial product. Yes, people can use it for free (right now) but ultimately they expect to sell access and profit from it. Also, no one else gets access to their training database, it is kept as some sort of trade secret.

for once, you are correct that this is not established law

Yay!

My perspective is that precedent supports training data for LLM’s as a fair use, and that strengthening copyright in the way proposed does not mitigate the harm being claimed by plaintiffs, and in fact increases harm to the greater public by gatekeeping access to automation tools and consolidating the benefits to already gigantic companies.

I wouldn't want to restrict or gatekeep access to art for genuine fair purpose uses. I agree with the Authors Guild rulings in those circumstances, I just disagree that LLM's are a similar enough circumstance that LLM's deserve the same exemption with how they're developed.

I really have to stress again that the issues and concerns being raised over AI cannot be sufficiently addressed through the use of copyright law.

I agree. Certainly, not copyright law as it exists right now, and even then there are so many aspects of the use of AI that fall well oustide the scope of copyright law.

Ultimately, my gripe is that a commercial business has used copyrighted work to develop a product without paying the rightsholders. Their product is their own unique creation, but the copyrighted work their product learned from was not. The training database they've used is not "research" because it is not scholarly; even if it were research, it is highly commercial in nature and as such does not warrant a fair use exemption.

[–] BreadstickNinja@lemmy.world 6 points 10 months ago (2 children)

Critical to understanding whether this applies is to understand "use" in the first place. I would argue it'd even more important because it's a threshold question in whether you even need to read 107.

17 U.S. Code § 106 - Exclusive rights in copyrighted works Subject to sections 107 through 122, the owner of copyright under this title has the exclusive rights to do and to authorize any of the following: (1)to reproduce the copyrighted work in copies or phonorecords; (2)to prepare derivative works based upon the copyrighted work; (3)to distribute copies or phonorecords of the copyrighted work to the public by sale or other transfer of ownership, or by rental, lease, or lending; (4)in the case of literary, musical, dramatic, and choreographic works, pantomimes, and motion pictures and other audiovisual works, to perform the copyrighted work publicly; (5)in the case of literary, musical, dramatic, and choreographic works, pantomimes, and pictorial, graphic, or sculptural works, including the individual images of a motion picture or other audiovisual work, to display the copyrighted work publicly; and (6)in the case of sound recordings, to perform the copyrighted work publicly by means of a digital audio transmission.

Copyright protects just what it sounds like- the right to "copy" or reproduce a work along the examples given above. It is not clear that use in training AI falls into any of these categories. The question mainly relates to items 1 and 2.

If you read through the court filings against OpenAI and Stability AI, much of the argument is based around trying to make a claim under case 1. If you put a model into an output loop you can get it to reproduce small sections of training data that include passages from copyrighted works, although of course nowhere near the full corpus can be retrieved because the model doesn't contain any thing close to a full data set - the models are much too small and that's also not how transformers architecture works. But in some cases, models can preserve and output brief sections of text or distorted images that appear highly similar to at least portions of training data. Even so, it's not clear that this is protected under copyright law because they are small snippets that are not substitutes for the original work, and don't affect the market for it.

Case 2 would be relevant if an LLM were classified as a derivative work. But LLMs are also not derivative works in the conventional definition, which is things like translated or abridged versions, or different musical arrangements in the case of music.

For these reasons, it is extremely unclear whether copyright protections are even invoked, becuase the nature of the use in model training does not clearly fall under any of the enumerated rights. This is not the first time this has happened, either - the DMCA of 1998 amended the Copyright Act of 1976 to add cases relating to online music distribution as the previous copyright definitions did not clearly address online filesharing.

There are a lot of strong opinions about the ethics of training models and many people are firm believers that either it should or shouldn't be allowed. But the legal question is much more hazy, because AI model training was not contemplated even in the DMCA. I'm watching these cases with interest because I don't think the law is at all settled here. My personal view is that an act of congress would be necessary to establish whether use of copyrighted works in training data, even for purposes of developing a commercial product, should be one of the enumerated protections of copyright. Under current law, I'm not certain that it is.

[–] archomrade@midwest.social 2 points 10 months ago

This is an extremely good write-up, thank you for this.

[–] TWeaK@lemm.ee -1 points 10 months ago (1 children)

(1)to reproduce the copyrighted work in copies or phonorecords

The works are copied in their entirey and reproduced in the training database. AI businesses do not deny this is copying, but instead claim it is research and thus has a fair use exemption.

I argue it is not research, but product development - and furthermore, unlike traditional R&D, it is not some prototype that is different and separate from the commercial product. The prototype is the commercial product.

(2)to prepare derivative works based upon the copyrighted work

AI can and has reproduced significant portions of copyrighted work, even in spite of the fact that the finished product allegedly does not include the work in its database (it just read the training database).

Furthermore, even if a human genuinely and honestly believes they're writing something original, that does not matter when they reproduce work that they have read before. What determines copyright infringement is the similarity of the two works.

If you read through the court filings against OpenAI and Stability AI, much of the argument is based around trying to make a claim under case 1.

The position that I take is that the arguments made against OpenAI and Stability AI in court are not complete. They're not quite good enough. However, that doesn't mean there isn't a valid argument that is good enough. I just hope we don't get a ruling in favour of AI businesses simply because the people challenging them didn't employ the right ammunition.

With regards to Case 2, I refer back to my comment about the similarity of the work. The argument isn't that the LLM itself is an infringement of copyright, but that the LLM, as designed by the business, infringes copyright in the same way a human would.

I definitely agree it is all extremely unclear. However, I maintain that the textual definition of the law absolutely still encompasses the feeling that peoples' work is being ripped off for a commercial venture. Because it is so commercial, original authors are being harmed as they will not see any benefit from the commercial profits.


I would also like to point you to my other comment, which I put a lot of time into and where I expanded on many other points (link to your instance's version): https://lemmy.world/comment/6706240

[–] archomrade@midwest.social 1 points 10 months ago (1 children)

The works are copied in their entirey and reproduced in the training database. AI businesses do not deny this is copying, but instead claim it is research and thus has a fair use exemption.

The copying of the data is not, by itself, infringement. It depends on the use and purpose of the copied data, and the defense argues that training a model against the data is fair use under TDM use-cases.

AI can and has reproduced significant portions of copyrighted work, even in spite of the fact that the finished product allegedly does not include the work in its database (it just read the training database).

The model does not have a 'database', it is a series of transform nodes weighted against unstructured data. The transformation of the copyrighted works into a weighted regression model is what is being argued is fair use.

Furthermore, even if a human genuinely and honestly believes they’re writing something original, that does not matter when they reproduce work that they have read before.

yup, and it isn't the act of that human reading a copyrighted work that is considered as infringement, it is the creation of the work that is substantially similar. In the same analogy, it wouldn't be the creation of the AI model that is the infringement, but each act of creation thereafter that is substantially similar to a copyrighted work. But this comes with a bunch of other problems for the plaintiffs, and would be a losing case without merit.

The position that I take is that the arguments made against OpenAI and Stability AI in court are not complete

The argument isn’t that the LLM itself is an infringement of copyright, but that the LLM, as designed by the business, infringes copyright in the same way a human would.

Trying really hard not to come off as rude, but there's a good reason why this isn't the argument being put forward in the lawsuits. If this was their argument, the LLM could be considered a commissioned agent, placing the liability on the agent commissioning the work (e.g. the human prompting the work) - not OpenAI or Stability - in much the same way a company is held responsible for the work produced by an employee.

I really do understand the anger and frustration apparent in these comments, but I would really like to encourage you to learn a bit more about the basis for these cases before spending substantial effort writing long responses.

[–] TWeaK@lemm.ee 1 points 10 months ago (1 children)

The copying of the data is not, by itself, infringement.

Copyright is absolute. The rightsholder has complete and total right to dictate how it is copied. Thus, any unauthorised copying is copyright infringement. However, fair use gives exemption to certain types of copying. The copyright is still being infringed, because the rightsholder's absolute rights are being circumvented, however the penalty is not awarded because of fair use.

This is all just pedantry, though, and has no practical significance. Saying "fair use means copyright has not been infringed" doesn't change anything.

it is a series of transform nodes weighted against unstructured data.

That's a database. Or perhaps rather some kind of 3D array - which could just be considered an advanced form of database. But yeah, you're right here, you win this pedantry round lol. 1-1.

it wouldn’t be the creation of the AI model that is the infringement, but each act of creation thereafter that is substantially similar to a copyrighted work. But this comes with a bunch of other problems for the plaintiffs, and would be a losing case without merit.

Yeah I don't want to go down the avenue of suing the AI itself for infringement. However...[^1][^2][^3]

Trying really hard not to come off as rude

You're not coming off as rude at all with what you've said, in fact I welcome and appreciate your rebuttals.

I really do understand the anger and frustration apparent in these comments, but I would really like to encourage you to learn a bit more about the basis for these cases before spending substantial effort writing long responses.

You say that as if I haven't enjoyed fleshing out the ideas and sharing them. By the way, right now I'm sharing with you lemmy's hidden citation feature :o)

Although, I was much happier replying to you before I just saw the downvotes you've apparently given me across the board. That's a bit poor behaviour on your part, you shouldn't downvote just because you disagree - and you can't even say that I'm wrong as a justification when the whole thing is being heavily debated and adjudicated over whether it is right or wrong.

I thought we were engaging in a positive manner, but apparently you've been spitting in my face.

[^1]: >but there's a good reason why this isn't the argument being put forward in the lawsuits.

[^2]: >the LLM could be considered a commissioned agent

[^3]: The LLM absolutely could be considered an agent, but the way it acts is merely prompted by the user. The actual behaviour is dictated by the organisation that built it. In any case, this is only my backup argument if you even consider the initial copying to be research - which it isn't.

[–] archomrade@midwest.social 2 points 10 months ago (1 children)

Copyright is absolute. The rightsholder has complete and total right to dictate how it is copied.

Really and truly, this is not how this works. The exemptions granted by the office of the registrar are granting an exemption to copyright claims against fair uses. It isn't talking about whether the claim can be awarded damages, it's talking about the claim being exempt in entirety. You can think about copyright as an exemption to the first amendment right to free speech, and the exemption to copyright as describing where that 'right' does not apply. Copyright holders do not get to control the use of their work where fair use has been determined by the registrar, which is reconsidered every 3 years.

This is all just pedantry, though, and has no practical significance. Saying “fair use means copyright has not been infringed” doesn’t change anything.

True enough, but it seems like it's important for your understanding in how copyright works.

Or perhaps rather some kind of 3D array - which could just be considered an advanced form of database. But yeah, you’re right here, you win this pedantry round lol. 1-1.

I wasn't being pedantic, that distinction is important for how copyright is conceptualized. The AI model is the thing being considered for infringement, so it's important to note that the works being claimed within it do not exist as such within the model. The '3-d array' does not contain copyrighted works. You can think of it as a large metadata file, describing how to construct language as analyzed through the training data. The nature and purpose of the 'work' is night-and-day different from the works being claimed, and 'database' is a clear misrepresentation (possibly even intentionally so) of what it is.

Yeah I don’t want to go down the avenue of suing the AI itself for infringement.

That was exactly what you pivoted to in your comment here, i'm not sure why you're now saying you don't want to go down that avenue. I'm confused what you're arguing at this point.

Although, I was much happier replying to you before I just saw the downvotes you’ve apparently given me across the board. That’s a bit poor behaviour on your part, you shouldn’t downvote just because you disagree - and you can’t even say that I’m wrong as a justification when the whole thing is being heavily debated and adjudicated over whether it is right or wrong.

I've down-voted your comments because they contain inaccuracies and could be misleading to others. You shouldn't let my grading of your comments reflect my attitude towards you; i'm sure you're a fine individual. Downvotes don't mean anything on Lemmy anyway, i'm not sure 'spitting in your face' is a fair or accurate description, but I don't want to invalidate your feelings, so I apologize for making you feel that way as that wasn't my intent.

[–] TWeaK@lemm.ee 1 points 10 months ago

I’ve down-voted your comments because they contain inaccuracies and could be misleading to others. You shouldn’t let my grading of your comments reflect my attitude towards you; i’m sure you’re a fine individual. Downvotes don’t mean anything on Lemmy anyway, i’m not sure ‘spitting in your face’ is a fair or accurate description, but I don’t want to invalidate your feelings, so I apologize for making you feel that way as that wasn’t my intent.

No worries, you've been very respectable. My feelings weren't particularly hurt, I just felt the need to call it out.

Personally, I'm against downvoting things merely because they are wrong. If someone says something that's wrong, it may well be a commonly held misconception, and downvoting it also demotes any correction that has been given, which means other people who hold the misconception are less likely to be corrected.

And that's beside the fact that I don't really think I'm completely wrong here :o)

That was exactly what you pivoted to in your comment here, i’m not sure why you’re now saying you don’t want to go down that avenue. I’m confused what you’re arguing at this point.

To be a little more specific, I don't want to go down the route of blaming AI itself for copyright infringement. That is to say, whether or not AI is bound by laws the way that humans are. I think it is only worthwhile considering whether the AI developer and/or the users are infringing copyright through their creation or use of AI. In particular, I think the legal or philosophical question of whether AI is affected by laws in the same way humans are is pointless when we're just talking about LLM's and not a true Artificial Intelligence.

The ‘3-d array’ does not contain copyrighted works. You can think of it as a large metadata file, describing how to construct language as analyzed through the training data. The nature and purpose of the ‘work’ is night-and-day different from the works being claimed, and ‘database’ is a clear misrepresentation (possibly even intentionally so) of what it is.

Yes absolutely, the LLM itself does not include copyrighted works. That's not what I'm arguing. The two issues I take are with the database of information the LLM is trained on. This database does contain copyrighted works, AI developers admit that it does, but they claim it is fair use research. I disagree with this claim, to use the terminology from one of your links their "research" is not "scholarly" - it is commercial product development.

The other issue is that the LLM can reproduce copyrighted work. While I agree with you in some sense that the user of the LLM is instructing it to infringe copyright, and thus the user is responsible, in another sense I think the developer is also responsible because they have given the tool the capability to do this. This is perhaps not a strong argument, particularly when the developers have made efforts to fix these "bugs" as they come to light.

However my most important point is that the developers have infringed copyright by building a training database full of copyrighted works, which the LLM was then trained on. The LLM itself isn't copyright infringement, but they infringed copyright to develop it.