andallthat

joined 1 year ago
[–] andallthat@lemmy.world 5 points 5 days ago* (last edited 5 days ago)

exactly! "I was energized to meet with the team in X and discuss our sales figures" or "congrats, company Y, for disrupting the market of foot creams" is the best use of AI.

I'm not sure how you would even be able to tell if that type of content is AI-generated or just plain old copy-pasted from one of a thousand similar posts

[–] andallthat@lemmy.world 23 points 6 days ago* (last edited 6 days ago) (5 children)

if you want to listen to 35 minutes of annoying music, press 1. If you want to get insulted personally by an operator stay on the line

[–] andallthat@lemmy.world 5 points 2 weeks ago

I don't think Musk would disagree with that definition and I bet he even likes it.

The key word here is "significant". That's the part that clearly matters to him, based on his actions. I don't care about the man and I don't think he's a genius, but he does not look stupid or delusional either.

Musk spreads disinformation very deliberately for the purpose of being significant. Just as his chatbot says.

[–] andallthat@lemmy.world 48 points 1 month ago (1 children)

I think I'm with him on this one. Replacing all the people on social with AI agents would give us back so much free time! And we could even restart socializing for real.

Go on Zuckerberg, give us a Facebook made only of AI agents creating fake pictures of inexistent gatherings and posting them, so other AIs can recommend them and million of other AIs can comment on them!

You are an unsung hero, Zuckerberg, but one day they'll understand and thank you

[–] andallthat@lemmy.world 4 points 1 month ago

on the other hand, when Putin's done killing off most of their own present and future workforce in a senseless war and completely tanking his own economy, that might be the equivalent of like $3

[–] andallthat@lemmy.world 15 points 1 month ago

Socials and the Internet in general would be a much better place if people stopped believing and blindly resharing everything they read, AI-generated or not.

[–] andallthat@lemmy.world 2 points 2 months ago* (last edited 2 months ago) (1 children)

I'm not sure we, as a society, are ready to trust ML models to do things that might affect lives. This is true for self-driving cars and I expect it to be even more true for medicine. In particular, we can't accept ML failures, even when they get to a point where they are statistically less likely than human errors.

I don't know if this is currently true or not, so please don't shoot me for this specific example, but IF we were to have reliable stats that everything else being equal, self-driving cars cause less accidents than humans, a machine error will always be weird and alien and harder for us to justify than a human one.

"He was drinking too much because his partner left him", "she was suffering from a health condition and had an episode while driving"... we have the illusion that we understand humans and (to an extent) that this understanding helps us predict who we can trust not to drive us to our death or not to misdiagnose some STI and have our genitals wither. But machines? Even if they were 20% more reliable than humans, how would we know which ones we can trust?

[–] andallthat@lemmy.world 14 points 2 months ago (1 children)

Most things to do with Green Energy. Don't get me wrong, I think solar panels or wind turbines are great. I just think that most of the reported figures are technically correct but chosen to give a misleadingly positive impression of the gains.

Relevant smbc: https://www.smbc-comics.com/comic/capacity

[–] andallthat@lemmy.world 4 points 4 months ago* (last edited 3 months ago)

I think they don't matter with outrage, because outrage explodes in ways that are hard to predict. I mean, I can see the problem with the ad now that it has been pointed out to me. After reading about it repeatedly, I now find it bad and ridiculous and what were they thinking? But at a first look, as a test audience I would have probably rated it as "meh, ok".

[–] andallthat@lemmy.world 6 points 4 months ago

AKA "shit, looks like now we need to re-hire some of those engineers"

[–] andallthat@lemmy.world 14 points 4 months ago (2 children)

TBH those same colleagues were probably just copy/pasting code from the first google result or stackoverflow answer, so arguably AI did make them more productive at what they do

view more: next ›