this post was submitted on 22 Feb 2024
488 points (96.2% liked)

Technology

59605 readers
3501 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Google apologizes for ‘missing the mark’ after Gemini generated racially diverse Nazis::Google says it’s aware of historically inaccurate results for its Gemini AI image generator, following criticism that it depicted historically white groups as people of color.

you are viewing a single comment's thread
view the rest of the comments
[–] intensely_human@lemm.ee 1 points 9 months ago (2 children)

What do you mean it has no world model? Of course it has a world model, composed of the relationships between words in language that describes that world.

If I ask it what happens when I drop a glass onto concrete, it tells me. That’s evidence of a world model.

[–] fidodo@lemmy.world 7 points 9 months ago* (last edited 9 months ago)

A simulation of the world that it runs to do reasoning. It doesn't simulate anything, it just takes a list of words and then produces the next word in that list. When you're trying to solve a problem, do you just think, well I saw these words so this word comes next? No, you imagine the problem and simulate it in both physical and abstract terms to come up with an answer.

[–] EpeeGnome@lemm.ee 3 points 9 months ago* (last edited 9 months ago) (1 children)

I can see the argument that it has a sort of world model, but one that is purely word relationships is a very shallow sort of model. When I am asked what happens when a glass is dropped onto concrete, I don't just think about what I've heard about those words and come up with a correlation, I can also think about my experiences with those materials and with falling things and reach a conclusion about how they will interact. That's the kind of world model it's missing. Material properties and interactions are well enough written about that it ~~simulates ~~ emulates doing this, but if you add a few details it can really throw it off. I asked Bing Copilot "What happens if you drop a glass of water on concrete?" and it went into excruciating detail about how the water will splash, mentions how it can absorb into it or affect uncured concrete, and now completely fails to notice that the glass itself will strike the concrete, instead describing the chemistry of how using "glass (such as from the glass of water)" as aggregate could affect the curing process. Having a purely statistical/linguistic world model leaves some pretty big holes in its "reasoning" process.

[–] fidodo@lemmy.world 2 points 9 months ago (1 children)

I believe you meant to say emulates instead of simulates

[–] EpeeGnome@lemm.ee 1 points 9 months ago

Thanks, that is a better word there.