this post was submitted on 01 Jun 2024
398 points (98.3% liked)

Technology

59589 readers
2910 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] SpaceNoodle@lemmy.world 9 points 5 months ago (5 children)

In what ways are you benefiting from a bevy of factually dubious query responses?

[–] brbposting@sh.itjust.works 10 points 5 months ago (1 children)

Can absolutely never blindly trust the hallucinating plagiarism machine.

It's great where either facts don't matter or you're personally in a position to vet all of its “factual” output 100%. Text revision, prompting for additional perspectives, prompting to challenge beliefs and identify gaps. Reformatting, quick and easy data extraction, outlining, brainstorming.

[–] SpaceNoodle@lemmy.world 2 points 5 months ago (1 children)

Reformatting and outlining as long as you go over and revise it again anyway, seemingly making that moot.

Data extraction as long as you don't care if the data is mangled.

Brainstorming is a good one, since off-the-wall ideas can be useful in that context.

[–] sugar_in_your_tea@sh.itjust.works 4 points 5 months ago (2 children)

In most cases I've seen AI used, the person spends as much time correcting it than they would if they just did the work without AI. So maybe it makes you feel more productive because a bunch of stuff happens all at once, but at least for text generation, I think it's more of a placebo.

[–] sudoreboot@slrpnk.net 3 points 5 months ago

It can at least get one unstuck, past an indecision paralysis, or give an outline of an idea. It can also be useful for searching though data.

[–] SpaceNoodle@lemmy.world -1 points 5 months ago (1 children)

If all I want is something blatantly false or legible yet nonsensical, like a modern lorem ipsum, it's a real time-saver.

[–] sugar_in_your_tea@sh.itjust.works 0 points 5 months ago (1 children)

Why not just use lorem ipsum? It's just a copy/paste, and without the liability of having false information if you forget to proofread it.

[–] SpaceNoodle@lemmy.world -2 points 5 months ago

I guess ChatGPT is just completely useless, then.

[–] Zos_Kia@lemmynsfw.com 7 points 5 months ago (1 children)

You cannot in all seriousness use a LLM as a research tool. That is explicitly not what it is useful for. A LLM's latent space is like a person's memory : sure there is some accurate data in there, but also a lot of "misremembered" or "misinterpreted" facts, and some bullshit.

Think of it like a reasoning engine. Provide it some data which you have researched yourself, and ask it to aggregate it, or summarize it, you'll get some great results. But asking it to "do the research for you" is plain stupid. If you're going to query a probabilistic machine for accurate information, you'd be better off rolling dice.

[–] SpaceNoodle@lemmy.world -2 points 5 months ago* (last edited 5 months ago) (1 children)

Exactly my point - except that the word "reasoning" is far too generous, as it implies that there would be some way for it to guarantee that its logic is sound, not just highly resembling legible text.

[–] Zos_Kia@lemmynsfw.com 0 points 5 months ago (1 children)

I don't understand. Have you ever worked an office job? Most humans have no way to guarantee their logic is sound yet they are the ones who do all of the reasoning on earth. Why would you have higher standards for a machine?

[–] SpaceNoodle@lemmy.world 3 points 5 months ago (1 children)

I have higher expectations for machines than humans, yes.

[–] Zos_Kia@lemmynsfw.com 0 points 5 months ago (1 children)

Sounds like a recipe for disappointment tbh. But on the other hand, sounds like you trust techno marketing a bit too much.

[–] SpaceNoodle@lemmy.world 0 points 5 months ago (1 children)

No, I just know how to spot the lies in a datasheet.

[–] Zos_Kia@lemmynsfw.com -1 points 5 months ago (1 children)

I"m not sure what lie and what datasheet you're referring to ?

[–] SpaceNoodle@lemmy.world 1 points 5 months ago

Just in general.

[–] kogasa@programming.dev 3 points 5 months ago (1 children)

I don't really query, but it's good enough at code generation to be occasionally useful. If it can spit out 100 lines of code that is generally reasonable, it's faster to adjust the generated code than to write it all from scratch. More generally, it's good for generating responses whose content and structure are easy to verify (like a question you already know the answer to), with the value being in the time saved rather than the content itself.

[–] SpaceNoodle@lemmy.world -2 points 5 months ago

It's good at regurgitating boilerplate, from what I've gathered.

[–] capital@lemmy.world 3 points 5 months ago* (last edited 5 months ago) (1 children)

This question betrays either your non-use or misuse of the products available. You're either just reading the headlines of the screw-ups or you're just bad at using the tool.

To directly answer your question:

  • Quick scripts in a variety of languages. Tested before being used on real data/systems.
  • Creating visual graphs of data in python and Jupyter notebooks with no prior knowledge of python itself or the tools it's running. In this case, I was able to update the way I wanted it to look in natural language, have it suggest code changes, and immediately try them in the notebook with great results.
  • Improving the sentiment of correspondence. Proofread before sending. It has better grammar and flow than a surprising number of correspondences I've come across at work. Sure, English may be their second language but it doesn't change the fact.
  • Quickly finding documentation pertaining to the query which, yes, you need to go read to verify any answers any LLM provides. Anyone using it regularly should know this by now.
  • Quick "do this in command line. What options are required" which is then immediately tested.
  • In one case, a news story was referenced in passing in a podcast I listen to. It stuck with me days later and I wanted to find actual articles written about it. I was able to describe what I was looking for in natural language and included as many details as I could remember and asked it to find articles for me. I found exactly what I was after.

But were you actually looking for a real response to your question?

[–] SpaceNoodle@lemmy.world -1 points 5 months ago (1 children)

It's worse at all programming tasks except boilerplate, especially with its tendency to inject booby traps. Not knowing how to use the programming language it emits becomes a significant problem.

Comparing a language model to an idiot is unfair to the idiot.

A normal search engine works for everything else.

Any well-defined query I've ever made of an LLM has resulted in hilariously bad results, but I suppose I was expecting it to do something that I couldn't already do better myself.

[–] capital@lemmy.world 4 points 5 months ago

I'm a systems administrator, not a programmer. Like I said, quick scripts. An LLM could probably parse my comment better than you, evidently.

Comparing a language model to an idiot is unfair to the idiot.

Oof.. Was this in reply to my bit about better grammar and ESL individuals?

A normal search engine works for everything else.

Fuck no. Especially the python visualization point.

Any well-defined query I’ve ever made of an LLM has resulted in hilariously bad results, but I suppose I was expecting it to do something that I couldn’t already do better myself.

I suppose you're just a god among men then. For the rest of us, it's useful and you've been given plenty of good answers to your disingenuous question.