this post was submitted on 05 Feb 2024
661 points (88.2% liked)
Memes
45727 readers
1034 users here now
Rules:
- Be civil and nice.
- Try not to excessively repost, as a rule of thumb, wait at least 2 months to do it if you have to.
founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
You're making a huge assumption; that an advanced parrot produces the same output as something with general intelligence. And I reject that assumption. Something with general intelligence can produce something novel. An advanced parrot can only repeat things it's already heard.
How do you define novel? Because LLMs absolutely have produced novel data.
LLMs can't produce anything without being prompted by a human. There's nothing intelligent about them. Imo it's an abuse of the word intelligence since they have exactly zero autonomy.
I use LLMs to create things no human has likely ever said and it's great at it, for example
'while juggling chainsaws atop a unicycle made of marshmallows, I pondered the existential implications of the colour blue on a pineapples dream of becoming a unicorn'
When I ask it to do the same using neologisms the output is even better, one of the words was exquimodal which I then asked for it to invent an etymology and it came up with one that combined excuistus and modial to define it as something beyond traditional measures which fits perfectly into the sentence it created.
You can't ask a parrot to invent words with meaning and use them in context, that's a step beyond repetition - of course it's not full dynamic self aware reasoning but it's certainly not being a parrot
Producing word salad really isn't that impressive. At least the art LLMs are somewhat impressive.
If you ask it to make up nonsense and it does it then you can't get angry lol. I normally use it to help analyse code or write sections of code, sometimes to teach me how certain functions or principles work - it's incredibly good at that, I do need to verify it's doing the right thing but I do that with my code too and I'm not always right either.
As a research tool it's great at taking a basic dumb description and pointing me to the right things to look for, especially for things with a lot of technical terms and obscure areas.
And yes they can occasionally make mistakes or invent things but if you ask properly and verify what you're told then it's pretty reliable, far more so than a lot of humans I know.