this post was submitted on 29 Apr 2024
195 points (94.9% liked)

Technology

59605 readers
3397 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] givesomefucks@lemmy.world 23 points 6 months ago (6 children)

If scientists made AI, then it wouldn't be an issue for AI to say "I don't know".

But capitalists are making it, and the last thing you want is it to tell an investor "I don't know". So you tell it to make up bullshit instead, and hope the investor believes it.

It's a terrible fucking way to go about things, but this is America...

[–] expr@programming.dev 39 points 6 months ago (2 children)

It's got nothing to do with capitalism. It's fundamentally a matter of people using it for things it's not actually good at, because ultimately it's just statistics. The words generated are based on a probability distribution derived from its (huge) training dataset. It has no understanding or knowledge. It's mimicry.

It's why it's incredibly stupid to try using it for the things people are trying to use it for, like as a source of information. It's a model of language, yet people act like it has actual insight or understanding.

[–] hatedbad@lemmy.sdf.org 1 points 6 months ago

you’re so close, just why exactly do you think people are using it for these things it’s not meant for?

because every company, every CEO, every VP, is pushing every sector of their companies to adopt AI no matter what.

most actual people understand the limitations you list, but it’s the capitalists at the table that are making AI show up where it’s not wanted

[–] VeganCheesecake@lemmy.blahaj.zone 28 points 6 months ago (3 children)

Uh, I understand the sentiment, but the model doesn't know anything. And it's legit really hard to differentiate between factual things and random bullshit it made up.

[–] DudeDudenson@lemmings.world 18 points 6 months ago (1 children)

Was gonna say, the AI doesn't make up or admit bullshit, its just a very advanced a prediction algorithm. It responds with what the combination of words that is most likely the expected answer.

Wether that is accurate or not is part of training it but you'll never get 100% accuracy to any query

[–] maynarkh@feddit.nl 1 points 6 months ago (3 children)

If it can name what the most likely combination is, couldn't it also know how likely that combination of words is?

[–] DudeDudenson@lemmings.world 7 points 6 months ago* (last edited 6 months ago)

It's not actually deciding anything, the AI thinking is marketing fluff really. But yes that's called confidence rating and it does. But at the scale of something like chatgpt that uses a snapshot of the entire internet and is non mutable there's no way to train it for every possible question. If you ask about a topic 99% of the internet gets wrong it'll respond the wrong thing with 99% confidence

[–] wahming@monyet.cc 3 points 6 months ago

No, because that requires it to understand the words. It doesn't.

[–] kent_eh@lemmy.ca 3 points 6 months ago

If it has been trained using questionable sources, or if it's training data includes sarcastic responses (without understanding that context), it isn't hard to imagine how confidently wrong some of the responses could be.

[–] Bishma@discuss.tchncs.de 8 points 6 months ago (1 children)

Yeah, no one can make it say "I don't know" because it is not really AI. Business bros decided to call it that and everyone smiled and nodded. LLMs are 1 small component (maybe) of AI. Maybe 1/80th of a true AI or AGI.

Honestly the most impressive part of LLMs is the tokenizer that breaks down the request, not the predictive text button masher that comes up with the response.

[–] Kichae@lemmy.ca 10 points 6 months ago

Honestly the most impressive part of LLMs is the tokenizer that breaks down the request, not the predictive text button masher that comes up with the response.

Yes, exactly! It's ability to parse the input is incredible. It's the thing that has that "wow" factor, and it feels downright magical.

Unfortunately, that also makes people intuitively trust its output.

[–] DarkThoughts@fedia.io 6 points 6 months ago* (last edited 6 months ago) (1 children)

This has nothing to do with scientists vs capitalists and everything with the fact that this is not actually "AI". Someone called it T9 (word prediction) on steroids and I find that much more fitting with how those LLMs work. It just mimics the way humans talk, but it doesn't actually converse intelligently or actually understands context - it just looks like it does, but only if you take it at face value and don't look deeper into it.

[–] howrar@lemmy.ca 2 points 6 months ago

It is made by scientists. And we don't know how to make the model determine whether or not it knows something. So far, we only have tools that tell us that something probably wasn't in the training set (e.g. using variance across models in a mixture of experts setup), but that doesn't tell us anything about how correct it is.

[–] set_secret@lemmy.world 0 points 6 months ago (1 children)

Just put this into GPT 4.

What's your view of the fizbang Raspberry blasters?

Gpt 'I'm not familiar with "fizbang Raspberry blasters." Could you provide more details or clarify what they are?'

It's a drink making machine from china

Gpt 'I don't have any specific information on the "fizbang Raspberry blasters" drink making machine. If it's a new or niche product, details might be limited online.'

So, in this instance is didn't hallucinate, i tried a few more made up things and it's consistent in saying it doesn't know of these.

Explanations?

[–] k110111@feddit.de 1 points 6 months ago (1 children)

Chatgpt and gpt4 are two different things. Gpt4 is like the engine and chatgpt is like a car. In early version they were pretty much the same thing, but nowadays they have implemented so much in chatgpt.

On top of that chatgpt4 is constantly trained for these scenarios, it is no longer a base model.

[–] set_secret@lemmy.world 1 points 6 months ago (1 children)

Oh ok thanks i thought this thread was about AI LLMs in general.

Weird i was downvoted for demonstrating the very thing that apparently (according to these very learned comments) AI can't do, actually doing it well. Seems like irrational bubble hate to me, common on reddit but getting more so on Lemmy it seems. "that guys asking topic based questions that make our comments look poorly thought out and potentially wrong, burn him"

[–] tonarinokanasan@lemmy.sdf.org 1 points 6 months ago* (last edited 6 months ago) (1 children)

This is a thing that is true of all LLMs, but it seems like you're misunderstanding the core issue. It CAN give outputs like that sometimes. What we CAN'T do is force it to give outputs like that ALL the time.

It will answer "I don't know" if its predictive text model guesses that the most common response to this would be "I don't know". To do that, to simplify a little, you could imagine that it reads your question, compares that to all the text in its training data, and tries to find the conversation that looks most like the question you asked, then answers whatever the person in the training data answered. But your exact question wasn't in its training data, so if you took that mental model, and instead had it compare to 1000 similar looking things in its training model and average them, then it would hopefully do a better job of coming up with something at least close to what you actually asked. Now take it to a million, or a billion.

When we're asking questions about the real world, we would prefer for it to answer based on knowledge about the real world. But what if it "matches" data from a work of fiction? Or just someone who doesn't know what they're talking about? Or true information, but about a different subject?

It doesn't know anything. It doesn't understand anything you say. It just looks at patterns that it learned from the training data and tries to guess what words are most likely to be said in that case. In other words, "here's one case where it didn't hallucinate" and "it will never hallucinate" are not the same thing at all.

Edit: To clarify, it doesn't search its training data to answer your question, so asking "was this in the training data" is impossible. By the time you interact with it, the data is long gone. It was just used for training.

[–] set_secret@lemmy.world 1 points 6 months ago (1 children)

Ok very long and detailed response, i was responding to the initial comments that explicitly said if you give ai a made up thing it will definitely hallucinate. Which i demonstrated to be false in (multiple times). I'm not suggesting it doesn't hallucinate a lot of the time still, but the comments were making out its 100% broken, and it clearly works for many queries very effectively, despite its limited applications. Im just suggesting we don't throw the baby out with the bathwater.

[–] tonarinokanasan@lemmy.sdf.org 1 points 6 months ago* (last edited 6 months ago) (1 children)

I think the trouble is, what baby are we throwing out with the bathwater in this case? We can't prevent LLMs from hallucinating (but we can mitigate it somewhat with carefully constructed prompts). So, use cases where we're okay with that are fair game, but any use case (or in this case, law?) that would require the LLM never hallucinates aren't attainable, and to get back earlier, this particular problem has nothing to do with capitalism.

[–] set_secret@lemmy.world 1 points 6 months ago
[–] Strobelt@lemmy.world -1 points 6 months ago

It is made by scientists. The problem is that said scientists are paid by investors mostly, or by grants that come from investors.