this post was submitted on 31 May 2024
347 points (98.3% liked)

Linux

48310 readers
645 users here now

From Wikipedia, the free encyclopedia

Linux is a family of open source Unix-like operating systems based on the Linux kernel, an operating system kernel first released on September 17, 1991 by Linus Torvalds. Linux is typically packaged in a Linux distribution (or distro for short).

Distributions include the Linux kernel and supporting system software and libraries, many of which are provided by the GNU Project. Many Linux distributions use the word "Linux" in their name, but the Free Software Foundation uses the name GNU/Linux to emphasize the importance of GNU software, causing some controversy.

Rules

Related Communities

Community icon by Alpár-Etele Méder, licensed under CC BY 3.0

founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] the_doktor@lemmy.zip 4 points 5 months ago (7 children)

I am just hoping governments will see the massive issues and copyright problems with AI and ban that garbage outright soon so all these companies eager to add their AI trash to every single product they ship will stop.

[–] sweng@programming.dev 8 points 5 months ago (6 children)

There is no general copyright issue with AIs. It completely depends on the training material (if even then), so it's not possible to make blanket statements like that. Banning technology, because a particular implementation is problematic, makes no sense.

[–] the_doktor@lemmy.zip 1 points 5 months ago (5 children)

The only relevant training material to make a truly complete dataset must include copyrighted material or you do not have a full set of data to draw from and thus it is useless. Stop defending this horrible technology.

[–] sweng@programming.dev 1 points 5 months ago* (last edited 5 months ago) (1 children)

What do you mean "full set if data"?

Obviously you can not train on 100% of material ever created, so you pick a subset. There is a a lot of permissively licensed content (e.g. Wikipedia) and content you can license (e.g. Reddit). While not sufficient for an advanced LLM, it certainly is for smaller models that do not need wide knowledge.

[–] the_doktor@lemmy.zip 1 points 5 months ago (1 children)

You can't even rely on Wikipedia to be right, and how is reddit "content you can license"? Its articles are owned by their sites, and the original stuff posted there is from the poster and is usually wildly inaccurate or outright wrong (or even downright dangerous). And even when they do pull in tons of stuff they shouldn't, the results are frequently laughably wrong.

You're not making a good argument for LLM crap here. Just accept the fact that it's a failed technology that needs to be shut down. Please. How are people so excited and gung-ho over this garbage, failed, laughably bad technology? It's almost like people WANT chaos.

[–] sweng@programming.dev 1 points 5 months ago (1 children)

Wikipedia is no less reliable than other content. There's even academic research about it (no, I will not dig for sources now, so feel free to not believe it). But factual correctness only matters for models that deal with facts: for e.g a translation model it does not matter.

Reddit has a massive amount of user-generated content it owns, e.g. comments. Again, the factual correctness only matters in some contexts, not all.

I'm not sure why you keep mentioning LLMs since that is not what is being discussed. Firefox has no plans to use some LLM to generate content where facts play an important role.

[–] the_doktor@lemmy.zip 1 points 5 months ago (1 children)

Sure hasn't helped AI/LLMs with accuracy yet. And never will. Computing doesn't actually think and reason, it's just mashing together bits of data it has, and if what it has now isn't accurate, how is anything going to be?

You and others continue to harp on how great this new technology is and meanwhile we have seen it do nothing but absolutely, laughably fail. You keep saying it will get better, but it won't. It is limited by the fact that computers don't work that way. Sick and tired of the people justifying this garbage "tech" that is stealing art, code, text, etc, sucking up huge amounts of power, and giving wrong information, telling people to do dangerous things and even kill themselves because computers don't know the difference.

Just admit it. AI/LLM is garbage. Please. Stop being a massive fanboy for something that has clearly, evidently, 100% failed miserably and dangerously.

[–] sweng@programming.dev 1 points 5 months ago

I think you are replying to the wrong person?

I did not say it helps with accuracy. I did not say LLMs will get better. I did not even say we should use LLMs.

But even if I did, non of your points are relevant for the Firefox usecase.

load more comments (3 replies)
load more comments (3 replies)
load more comments (3 replies)