this post was submitted on 13 Jan 2024
919 points (98.7% liked)

Technology

59674 readers
3235 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

The Pentagon has its eye on the leading AI company, which this week softened its ban on military use.

you are viewing a single comment's thread
view the rest of the comments
[–] funkforager@sh.itjust.works 279 points 10 months ago (56 children)

Remember when open ai was a nonprofit first and foremost, and we were supposed to trust they would make AI for good and not evil? Feels like it was only Thanksgiving…

[–] Dave@lemmy.nz 129 points 10 months ago (3 children)

I mean, there was all that drama where the board formed to prevent this from happening kicked out the CEO trying to do this stuff, then the board got booted out and replaced with a new board and brought back that CEO guy. So this was pretty much going to happen.

[–] hoshikarakitaridia@sh.itjust.works 68 points 10 months ago (1 children)

And some people pointed it out even back then. There were signs that the employees were very loyal to Altmann, but Altmann didn't meet the security concerns of the board. So stuff like this was just a matter of time.

[–] deweydecibel@lemmy.world 33 points 10 months ago (1 children)

People pointed this out as a point in Altmann's favor, too. "All the employees support him and want him back, he can't be a bad guy!"

Well, ya know what, I'm usually the last person to ever talk shit about the workers, but in this case, I feel like this isn't a good thing. I sincerely doubt the employees of that company that backed Altmann had taken any of the ethics of the tool they're creating into account. They're all career minded, they helped develop a tool that is going to make them a lot of money, and I guarantee the culture around that place is futurist as fuck. Altmann's removal put their future at risk. Of course they wanted him back.

And frankly I don't think you can spend years of your life building something like ChatGBT without having drunk the Koolaid yourself.

The truth is OpenAI, as a body, set out to make a deeply destructive tool, and the incentives are far, far too strong and numerous. Capitalism is corrosive to ethics; it has to be in enforced by a neutral regulatory body.

[–] SuckMyWang@lemmy.world 3 points 10 months ago (1 children)

The engineers are likely seeing this from an arms race point of view. Possibly something like the development of an a-bomb where it’s a race against nations and these people at the leading edge can see things we cannot. While money and capitalistic factors are at play, foreseeing your own possible destruction or demise by not being ahead of the game compared to china may be a motivating factor too.

[–] guacupado@lemmy.world 2 points 10 months ago

Bless your heart, sweet summer child.

[–] Sasha@lemmy.blahaj.zone 38 points 10 months ago (3 children)

Effective altruism is just capitalism camoflauge, it's also just really bad at being camoflauge

[–] iAvicenna@lemmy.world 19 points 10 months ago

helps you get a lot of community support and publicity during startup and then you don't have to give a damn about them once you take off

[–] Knock_Knock_Lemmy_In@lemmy.world 11 points 10 months ago

Effective altruism could work if the calculation of "amount of good" an action creates wasn't performed by the person performing that action.

E.g. I feel I'm doing a lot of good buying this $30m penthouse in the Bahamas.

[–] littlebluespark@lemmy.world 0 points 10 months ago (1 children)

You had two chances to spell camouflage correctly and you missed twice? I mean. Points for consistency, at least? 🤪

[–] Sasha@lemmy.blahaj.zone 4 points 10 months ago (2 children)

I can't spell, don't blame me for relying on an ordinarily quite useful tool.

[–] littlebluespark@lemmy.world 1 points 10 months ago

No judgement, autocorrect is my damn nemesis. 🤗🤘🏼

[–] afraid_of_zombies@lemmy.world 1 points 10 months ago (1 children)

Did they kick the CEO out for doing this or was it because of something else?

[–] Dave@lemmy.nz 6 points 10 months ago (1 children)

This summary article says the board stated:

"Mr. Altman’s departure follows a deliberative review process by the board, which concluded that he was not consistently candid in his communications with the board, hindering its ability to exercise its responsibilities," OpenAI's post said. "The board no longer has confidence in his ability to continue leading OpenAI."

The article also says:

Rumors and speculation swirled on social media, with tech industry heads, reporters, and onlookers trying to make sense of the situation based on what little information was provided in the board's announcement. Tech journalist Kara Swisher quickly reported that based on what information she had from sources, there was a "misalignment" between OpenAI's for-profit side, represented by Altman, and the nonprofit side, which is controlled by the board.

As far as I know the exact issue was not made public, but basically the board is there to make sure the company puts ethics over profits. Altman was hiding stuff from the board (presumably because they would consider it in conflict with their goal), and so the board fired him. But then there was an uproar from the investors, Microsoft almost ended up hiring half the company as they threatened to resign in droves, and in the end the board resigned and was replaced.

Does that answer the question?

load more comments (52 replies)