this post was submitted on 08 Apr 2026
254 points (97.0% liked)
Technology
83632 readers
4162 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
The point is, the LLM is not 'lying' to you. It's showing you information. It doesn't 'know' whether the information is true or not. It also doesn't 'care'. Because it is a statistical model and is incapable of those things. And if you scroll back to my initial point, I said "technically, it's not lying, because lying requires intent to deceive, and LLMs don't have intent"
What's the point of making this semantic difference though?
Because 1) it's true and the article is a bit misleading as to whom is actually doing the lying and 2) it's important to remember that LLM are not sentient and to push back against the tide of language which subtly suggests they are.