this post was submitted on 05 Apr 2024
521 points (96.6% liked)

Technology

59534 readers
3209 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] just_change_it@lemmy.world 66 points 7 months ago* (last edited 7 months ago) (6 children)

Jesus christ these headlines mislead everything.

They were using machine learning to try and figure out what people were buying. Machine learning has lots of errors until you train it. The "hundreds of workers" were training it by telling it what each thing was. E.g. it was creating training data for it to learn from.

The goal was to train ML enough so that humans were rarely necessary, obviously.

[–] baru@lemmy.world 64 points 7 months ago (1 children)

Jesus christ these headlines mislead everything.

One article included how often employees needed to look at the cameras. That was the case in something like 80% of the times people went in to shop.

The goal was to train ML enough so that humans were rarely necessary, obviously.

The headline is pretty accurate. That might have been the goal, but they didn't come close. And now they are closing down those stores.

Seems that they utterly failed in the goal.

Machine learning has lots of errors until you train it.

These stores were open for a pretty long time. It's not a given that it's just a matter of training.

[–] BakerBagel@midwest.social 26 points 7 months ago (1 children)

Except they still had thousands if employees in India watching the surveillance tapes to see what people bought and charged them for it

Amazon can claim this was a stop gap all they want, but the truth is that the technology behind the core concept isn't there and they just pretended it worked so the project head wouldn't have to explain why they are behind schedule and over budget. It's the same as with their drone delivery service 10 years ago. All smoke and mirrors to make moron tech bros cream themselves

[–] echodot@feddit.uk 1 points 7 months ago (1 children)

You need training data though I don't understand what the problem is. Hell it doesn't even matter if they never actually make the technology work that'll be their problem. They can't lie and tell people it works if it doesn't but as far as I'm aware they're not actually doing that.

I don't like amazing very much but they do enough crappy things for you to actually get upset about so it just seems odd that you would pick this hill to die on.

[–] BakerBagel@midwest.social 9 points 7 months ago

They gave up collecting that data though, that's why they are shutting down the department. All they managed to do was outsource grocery store cashiers to India, which seems like an exceptional shitty thing to me

[–] Zron@lemmy.world 24 points 7 months ago (1 children)

My goal is to build a fusion reactor.

I will hire Indian call center workers to add fuel to my diesel generator until the fusion is up and running.

This plan makes sense to certain people on the internet.

[–] AnUnusualRelic@lemmy.world 4 points 7 months ago

You could probably get lots of funding with this. I say go for it.

[–] MajorHavoc@programming.dev 22 points 7 months ago

The goal was to train ML enough so that humans were rarely necessary, obviously.

Yes, that's the goal.

There's a long rich history of AI like outcomes being mimiced by just hiding the human who does the work. That's actually the source of the name of Amazon's own "Mechanical Turk" service.

Not being actively watched by an army of underpaid workers is effectively still on the "someday...maybe" feature list for this thing, unless Amazon (famous for making delivery workers pee in soda bottles, and allowing warehouse workers to get heat stroke) somehow provides credible proof that they've actually grown past that.

I, as someone with substantial professional ML experience, won't take Amazon at their word, when they claim the ML has alleviated the need for the army of workers watching cameras. That's bullshit marketing promise, until proven otherwise. Particularly coming from Amazon.

Moving away from the people watching to using pure AI is well within the realm of possibility.

But good AI maintainers cost more per hour to pay than the entire army of mechanical turk "trainers". So I am skeptical of any claim that Amazon, in particular, has done the right thing here.

So it's very fair to assume you're being watched in one of those stores, until real credible evidence is provided that you're not.

[–] TurtleJoe@lemmy.world 19 points 7 months ago

They were using machine learning to try and figure out what people were buying. Machine learning has lots of errors ~~until you train it.~~

Machine Learning, no matter how well trained or advanced, is just doing a make-em-up.

Besides that, in this case the experiment has been going on for years and humans were still doing like 70% of the work. It was a failure, that's why Amazon shut it down

[–] steeznson@lemmy.world 6 points 7 months ago (1 children)

That's their excuse but it is convenient for them that in order to train the AI the workers need to follow the exact same steps as what an AI would be doing if it was sufficiently trained. We can't say as outsiders to what extent the actual work is assisted by AI. Seems likely that it is largely a manual process.

[–] Railcar8095@lemm.ee 0 points 7 months ago

I understand the spirit, but that's how it goes. You have somebody doing the work, as you want the ML to do it, and then feed the data. It's the same when they get oncology scans that have been diagnosed by well paid doctors, somebody who knows does and the machine tries to replicate.

What very likely happened is that the failure rate platoed much higher than they expected, and all this time the goal was to lower it. Remember, it's cheaper to have 0 people in India than 1, specially with AWS in mind.

Moreover, even if the accuracy was incredibly high, they would still need people reviewing. You have to review random events to ensure the model keeps performing well and to evaluate the ones with low confidence or suspicious.