this post was submitted on 16 Feb 2024
265 points (100.0% liked)

Technology

59605 readers
3175 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Air Canada must pay damages after chatbot lies to grieving passenger about discount | Airline tried arguing virtual assistant was solely responsible for its own actions::Airline tried arguing virtual assistant was solely responsible for its own actions

top 31 comments
sorted by: hot top controversial new old
[–] tiramichu@lemm.ee 127 points 9 months ago (1 children)

Shame on Air Canada for even fighting it.

I'm glad for this ruling. We need to set a legal precedent that chatbots act on behalf of the company. And if businesses try to claim that chatbots sometimes make mistakes then too bad - so do human agents, and when this happens in this customer's favour it needs to be honoured.

Companies want to use AI to supplement and replace human agents, but without any of the legal consequences of real people. We cannot let them have their cake and eat it at the same time.

[–] NocturnalEngineer@lemmy.world 28 points 9 months ago (1 children)

If it was a human agent, surely they would still liable?

They're an agent of the company. They're acting on behalf of the company, in accordance to their policy and procedures. It then becomes a training issue if they were providing incorrect information?

[–] tiramichu@lemm.ee 33 points 9 months ago

Yes, if it was a human agent they would certainly be liable for the mistake, and the law very much already recognises that.

That's my whole point here; the company should be equally liable for the behaviour of an AI agent as they are for the behaviour of a human agent when it gives plausible but wrong information.

[–] elvis_depresley@sh.itjust.works 72 points 9 months ago (1 children)

I kinda agree with this. If companies are going to replace human support (phone, chat or in person) with an LLM to save costs, then they should live with the consequences.

[–] thanks_shakey_snake@lemmy.ca 67 points 9 months ago (1 children)

Only kinda? To me, the "we're not liable because we have no idea how this technology is going to behave" argument is very unambiguously not acceptable.

[–] Modva@lemmy.world 66 points 9 months ago (1 children)

Motherfuckers tried to get away from responsibility for their own systems?

Air Canada, disgusting.

[–] laughterlaughter@lemmy.world 35 points 9 months ago

"Our Air Canada flight attendant punched you in the face for no reason, then our CEO kicked you in the nuts? Not our fault. They're independent agents."

[–] asteriskeverything@lemmy.world 47 points 9 months ago (1 children)

In effect, Air Canada suggests the chatbot is a separate legal entity that is responsible for its own actions

This is some corporations are people bs they are trying to get away with. This wasn't about greed over a couple hundred bucks, it's about precedent and boy were they trying to set a harmful one for the consumer in ANY industry that utilizes AI with customer support, perhaps other applications as well.

[–] TheFriar@lemm.ee 19 points 9 months ago (1 children)

So not only are they making the user experience way, way worse, they are trying to cut all costs and shovel them off onto us. I don’t remember where I read it, probably here a couple weeks ago, but I read and article stating how companies use the internet backwards. Instead of the internet being a tool for its customers, companies use it as a tool to protect themselves from the customers. We are filtered through purposefully aggravating automated call systems, or put through Chatbots as a measure to simplify us.

Is anyone down for a fucking Revolution against this insanely backwards concept of modern life? I am.

[–] asteriskeverything@lemmy.world 1 points 9 months ago (1 children)

That sounds fascinating, if you come across it again please send the link my way!

[–] TheFriar@lemm.ee 2 points 9 months ago

I’ve been trying tofind it since your comment and I just can’t place where I came across it! It was an opinion piece on some low rent looking leftist blog, so it’s hard to search for. I will keep looking though, because I’d like to find it again and I’ll share it when I do.

[–] thanks_shakey_snake@lemmy.ca 36 points 9 months ago (1 children)

Hundreds of dollars were on the line in this case! HUNDREDS!

[–] tiramichu@lemm.ee 28 points 9 months ago* (last edited 9 months ago) (2 children)

Hundreds in this case, but millions in the long term.

I can see why Air Canada wanted to fight it, because if they accept liability it sets a precedent that they should also accept liability for similar cases in future.

And they SHOULD accept liability, so I'm glad Air Canada lost and were forced to!

[–] brsrklf@jlai.lu 11 points 9 months ago (1 children)

The solution would be easy, just stop having an LLM chatbot.

But I suspect they don't want to because someone sold them on how good and cheap and human-resource-free it was, and now they think they're too invested.

[–] thanks_shakey_snake@lemmy.ca 2 points 9 months ago (1 children)

Plus just the general sentiment that you're not businessing right if you don't something something AI.

[–] rottingleaf@lemmy.zip 2 points 9 months ago

Plus just the general sentiment that you’re not businessing right if you don’t something something AI.

Feel my blood boiling at the very thought of people choosing to use something buzzwordy like blockchain or "AI", despite likely no competent person advising them to employ that, AND then trying to clean themselves of the responsibility when it misfires.

That's as if drunk driving leading to car crash was blamed on the air, because "having fun is not a crime".

Only with computing these people unironically think that nobody should be responsible, because everybody they respect is as clueless as themselves, so "nobody knows how it works, it's a frontier, see".

[–] thanks_shakey_snake@lemmy.ca 2 points 9 months ago

Yeah great point. Or similar cases in the past, even.

[–] Rentlar@lemmy.ca 22 points 9 months ago* (last edited 9 months ago)

https://lemmy.ca/post/15412680

See above for a !Canada@lemmy.ca article/comment section on this topic.

Air Canada is well known for the "how can we give less service for more money" mindset, and every other Canadian airline is better (WestJet is trying to compete for last place though).

If you use chatbots as your customer support agents, then you have to be responsible for any reasonable decision it makes. If you didn't "train it" properly like you would a new hire then that's on you.

[–] autotldr@lemmings.world 11 points 9 months ago

This is the best summary I could come up with:


Air Canada must pay a passenger hundreds of dollars in damages after its online chatbot gave the guy wrong information before he booked a flight.

Jake Moffatt took the airline to a small-claims tribunal after the biz refused to refund him for flights he booked from Vancouver to Toronto following the death of his grandmother in November last year.

Before he bought the tickets, he researched Air Canada's bereavement fares – special low rates for those traveling due to the loss of an immediate family member – by querying its website chatbot.

Unhappy with this situation – a support bot telling him the wrong info – Moffatt took the airline to a tribunal, claiming the corporation was negligent and misrepresented information, leaving him out of pocket.

Air Canada, however, argued it shouldn't be held liable for the chatbot's faulty outputs, without explaining why, which baffled tribunal member Christopher Rivers.

Air Canada said its chatbot provided a link to a page on its website explaining that refunds for discounted fares cannot be claimed retroactively, and Moffatt should have clicked on it.


The original article contains 626 words, the summary contains 180 words. Saved 71%. I'm a bot and I'm open source!