this post was submitted on 07 Aug 2024
129 points (93.3% liked)

Technology

59569 readers
3825 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] MentalEdge@sopuli.xyz 10 points 3 months ago* (last edited 3 months ago) (1 children)

And to have conversation, behind the scenes, each prompt gets the entire conversation so far tacked on.

The model itself is static, it doesn't work like a brain that changes in response to stimulus, or form memories.

To converse about something, the entirety of an exchange is fed back into the model all over again each time it needs to produce a response. In fact, this can happen several times over for each word in a response.

It's basically an attempt at duct-taping the ability to form memories onto an otherwise static system. It works, but I don't see how that way of doing it could ever land LLMs in the land of real consciousness.

It basically means these models "think" in frames, but each frame gets exponentially heavier to process, as it has to ingest every frame that came before.

[–] mozz@mbin.grits.dev 3 points 3 months ago

OpenAI at least is now attempting to bolt on a “memory” by having the LLM spit out short snippets of what it might need to know later, which it then has access to when completing later prompts. Like everything else post-GPT-4, it seems fine but doesn’t work really all that well at what it is intended to do.