this post was submitted on 12 Sep 2024
187 points (88.2% liked)
Technology
59495 readers
3114 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I just love how people seem to want to avoid using the word lie.
It’s either misinformation, or alternative facts, or hallucinations.
Granted, a lie does tend to have intent behind it, so with ChatGPT, it’s probably better to say falsehood, instead. But either way, it’s not fact, it’s not truth, and people, especially schools, should stop using it as a credible source.
Being wrong is not the same as lying. When LLMs start giving wrong answers on purpose to mislead people we would have a big problem.
The thought of a maliciously deceptive AGI is terrifying to me. Many, many people will trust it until it's too late.
There was a recent paper that argues 'bullshitting' is the most apt analogy. I.e. telling something to satisfy the other person without caring about the truth content of what you say
What about thw term "incorrect facts"?