this post was submitted on 30 Jul 2024
959 points (97.9% liked)
Technology
59963 readers
3387 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Learning to use the tools available to you is not "fake" it's being smart. Anyone who would be like "oh you recognize your weak point and have found and used a tool effectively to minimize it...you're fired/get out of my life" is an asshole and an idiot.
If you use binggpt as a translator tool, and put a disclaimer that these are not your own words - kudos, you removed the need for a translator and the latency associated.
However, if you claim that you speak English and use this tool to create a false impression of proficiency, that is just usual lying.
Furthermore, lacking proficiency in any language and using a tool to "beautify" a paragraph in said language will generally fail to improve communication, because chatgpt is trying to infer and add information which just isnt there (details, connotations, phraseologisms). Will just add more garbage to the conversation, and most likely words and meanings that just arent yours.
It's fine. Eventually when people start using this crap en masse the people on the other end will just be using LLMs to distill the bullshit down to 3 key points anyway.
That would be bizarre, lol
Let’s say one person writes 3 pages with some key points, then another extracts modified points due to added llm garbage then sends them again in 2 page essay to someone else and they again extract modified points. Original message was long gone and failure to communicate occurred but bots talk to each other so to say further producing even more garbage
In the end we are drowning in humongous pile of generated garbage and no one can effectively communicate anymore
The funny thing is this is mostly true without LLMs or other bots. People and institutions cant communicate because of leviathan amounts of legalese, say-literally-nothing-but-hide-it-in-a-mountain-of-bullshitese, barely-a-correlation-but-inflate-it-to-be-groundbreaking-ese, literally-lie-but-its-too-complicatedly-phrased-nobody-can-call-false-advertising-ese.
What about using an LLM to extract actual EULA key points?
I wouldn’t rely on LLM to read anything for you that matters. Maybe it will do ok nine out of ten times but when it fails you won’t even know until it is too late.
What if Eula itself was chat gpt generated from another chat generated output from another etc.. madness. Such Eula will be pure garbage suddenly and cutting costs no one will even notice relying on ai so much until it’s all fubar
So sure it will initially seem like a helpful tool, make key points from this text that was generated by someone from some other key points extracted by gpt but the mistakes will multiply in each iteration.