this post was submitted on 10 Jun 2025
86 points (75.9% liked)

Technology

72739 readers
1532 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] givesomefucks@lemmy.world 3 points 1 month ago (7 children)

Although the chatbot had been given a "baseline board" to learn the game and identify pieces, it kept mixing up rooks and bishops, misread moves, and "repeatedly lost track" of where its pieces were. To make matters worse, as Caruso explained, ChatGPT also blamed Atari's icons for being "too abstract to recognize" — but when he switched the game over to standard notation, it didn't perform any better.

For an hour-and-a-half, ChatGPT "made enough blunders to get laughed out of a 3rd grade chess club" while insisting over and over again that it would win "if we just started over," Caruso noted. (And yes, it's kind of creepy that the chatbot apparently referred to itself and the human it was interfacing with as "we.")

It's fucking insane it couldn't keep track of a board...

And it's concerning how confident it is that it will work, because the idiots asking it stuff will believe it. It'll keep failing and keep saying next time will work, because it's built to maximize engagement.

[–] pennomi@lemmy.world 4 points 1 month ago (6 children)

Spatial reasoning has always been a weakness of LLMs. Other symptoms include the inability to count and no concept of object permanence.

[–] givesomefucks@lemmy.world -3 points 1 month ago (5 children)

Yeah, but it's chess...

The LLM doesn't have to imagine a board, if you feed it the rules of chess and the dimensions of the board it should be able to "play in its head".

For a human to have that kind of working memory would be a genius level intellect and years of practice at the game.

But human working memory is shit compared to virtually every other animal. This and processing speed is supposed to be AI's main draw.

[–] Rhaedas@fedia.io 3 points 1 month ago

LLMs can be good at openings. Not because it is thinking through the rules or planning strategies, but because opening moves are likely in most general training data from various sources. It's copying the most probable reaction to your move, based on lots of documentation. This can of course break down when you stray from a typical play style, as it has less to choose from in the options of probability, and only a few moves in there won't be any more since there's a huge number of possible moves.

I.e., there's no calculations involved. When you play a LLM at chess, you're playing a list of common moves in history.

An even simpler example would be to tell the LLM that its last move was illegal. Even knowing the rules you just told it, it will agree and take it back. This comes from being trained to give satisfying replies to a human prompt.

load more comments (4 replies)
load more comments (4 replies)
load more comments (4 replies)