this post was submitted on 27 Apr 2026
1129 points (98.6% liked)

Technology

84199 readers
3253 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] DarkSurferZA@lemmy.world 2 points 4 hours ago

Never f**king guess my dude

[–] IronKrill@lemmy.ca 33 points 9 hours ago (1 children)

The AI agent was set to complete a routine task in the PocketOS staging environment. However, it came up against a barrier “and decided — entirely on its own initiative — to 'fix' the problem by deleting a Railway volume,” writes Crane, as he starts to describe the difficult-to-believe series of unfortunate events.

Quite easy-to-believe, really.

These multiple safeguards toppling in rapid succession

Multiple safeguards? Really? Multiple paragraph prompts are not multiple safeguards... it's half a safeguard at best. Applying limits on what the AI can do is a safeguard.

[–] Zizzy@lemmy.blahaj.zone 26 points 9 hours ago (1 children)

These people think giving the genai a prompt is coding. They dont understand the difference between actually coding in limits and just writing "pretty please dont delete everything"

[–] aesthelete@lemmy.world 14 points 8 hours ago (2 children)

I'm shocked and appalled that my addition of "do NOT make any mistakes!" didn't singlehandedly make the word guessing technology underneath perfect.

[–] MadhuGururajan@programming.dev 4 points 5 hours ago

Lol this is just like saying "I do declare bankruptcy"

[–] korazail@lemmy.myserv.one 2 points 6 hours ago

Who could have predicted this!?

Not an LLM, that's for sure. Maybe all the people screaming about this exact scenario, though.

[–] realitista@lemmus.org 9 points 7 hours ago

Can you get an AI to code? Yes. Can you get it to stop you from running your operation in such a stupid way that it will end up destroying it? No.

[–] subnormal@lemmy.dbzer0.com 22 points 9 hours ago

Reminder that Anthropic's AI system was used in targeting the school in Minab, killing 120 students. https://www.washingtonpost.com/national-security/2026/03/11/us-strike-iran-elementary-school-ai-target-list/

The company is suing to be able to supply the US military again. It is in bed with the fascists.

[–] PerogiBoi@lemmy.ca 30 points 12 hours ago

That's great to hear.

[–] fum@lemmy.world 35 points 17 hours ago (23 children)

This is absolutely hilarious. "AI" users getting what they deserve chef's kiss

[–] SaveTheTuaHawk@lemmy.ca 2 points 4 hours ago

This is what happens when there is a new technology and companies are run by commerce grads, not scientist or engineers that understand the technology.

load more comments (22 replies)
[–] captcha_incorrect@lemmy.world 15 points 16 hours ago (2 children)

This was on Hacker News: https://news.ycombinator.com/item?id=47911524

Twitter link: https://xcancel.com/lifeof_jer/status/2048103471019434248

Hacker New's sentiment on this from the comments I've read is that it is the author's own fault.

[–] UnrepententProcrastinator@lemmy.ca 11 points 12 hours ago (1 children)

As much as I want to blame AI for this, there are many hurdles for the user to get through to even allow Claude to do that. I'd be very suprised if that's not user error.

load more comments (1 replies)
load more comments
view more: next ›