this post was submitted on 17 Feb 2024
162 points (96.6% liked)

World News

39517 readers
2963 users here now

A community for discussing events around the World

Rules:

Similarly, if you see posts along these lines, do not engage. Report them, block them, and live a happier life than they do. We see too many slapfights that boil down to "Mom! He's bugging me!" and "I'm not touching you!" Going forward, slapfights will result in removed comments and temp bans to cool off.

We ask that the users report any comment or post that violate the rules, to use critical thinking when reading, posting or commenting. Users that post off-topic spam, advocate violence, have multiple comments or posts removed, weaponize reports or violate the code of conduct will be banned.

All posts and comments will be reviewed on a case-by-case basis. This means that some content that violates the rules may be allowed, while other content that does not violate the rules may be removed. The moderators retain the right to remove any content and ban users.


Lemmy World Partners

News [email protected]

Politics [email protected]

World Politics [email protected]


Recommendations

For Firefox users, there is media bias / propaganda / fact check plugin.

https://addons.mozilla.org/en-US/firefox/addon/media-bias-fact-check/

founded 2 years ago
MODERATORS
 

Company claimed its chatbot ‘was responsible for its own actions’ when giving wrong information about bereavement fare

Canada’s largest airline has been ordered to pay compensation after its chatbot gave a customer inaccurate information, misleading him into buying a full-price ticket.

Air Canada came under further criticism for later attempting to distance itself from the error by claiming that the bot was “responsible for its own actions”.

Amid a broader push by companies to automate services, the case – the first of its kind in Canada – raises questions about the level of oversight companies have over the chat tools.

all 14 comments
sorted by: hot top controversial new old
[–] Nurse_Robot 47 points 10 months ago (2 children)

"we're not responsible for the actions of our employees, even when we literally programmed them"

What a fucking joke, it's like they're saying the quiet part out loud

[–] Deestan 16 points 10 months ago (2 children)

It sounds like they are leaning on the chatbot being recognized as a sentient being. That's a pretty long shot.

[–] SlopppyEngineer 14 points 10 months ago

In that case the chat bot is an employee and must receive a wage or this is considered slavery.

[–] Nurse_Robot -1 points 10 months ago

Next Gen enshitification

[–] LufyCZ -2 points 10 months ago (1 children)

Well it's true, to a certain extent.

If an employee (or a chatbot, for that matter), promised an egregious sum for no reason, I don't think the company should be liable either.

Imagine getting hired to do support, having a friend open a chat and you promising to give him a milion dollars. Makes no sense.

But getting mislead about ticket pricing and them then refusing to refund the fare at least partially (the part that they promised would not be charged) is absolutely something they should be liable for.

And lawyer fees plus some pocket money for wasting peoples' time, if getting a refund entails more than an email or two.

[–] [email protected] 4 points 10 months ago

Yes, it was reasonable to believe was the point. I think what’s also interesting is the bot referred them to the correct information, which was part of the defence. However, the ruling said that both were provided by the company, the customer had no reason to believe the website gave more accurate information in one place compared to another.

[–] [email protected] 27 points 10 months ago

This chatbot no longer works for the airline according to its LinkedIn profile.

[–] [email protected] 16 points 10 months ago

"separate legal entity"

That the airline completely controls and has since updated the chatbot's programming.

[–] SpaceNoodle 12 points 10 months ago

And not even any punitive damages. You know they've absolutely screwed over others with this.

[–] [email protected] 4 points 10 months ago

Air Canada sucks. I avoid them as much as possible.

[–] inb4_FoundTheVegan 4 points 10 months ago

Good. Let's call it bug testing as we see what sort of deals various chatbots will agree too!

[–] [email protected] 2 points 10 months ago

This is the best summary I could come up with:


Canada’s largest airline has been ordered to pay compensation after its chatbot gave a customer inaccurate information, misleading him into buying a full-price ticket.

In 2022, Jake Moffatt contacted Air Canada to determine which documents were needed to qualify for a bereavement fare, and if refunds could be granted retroactively.

According to Moffat’s screenshot of a conversation with the chatbot, the British Columbia resident was told he could apply for the refund “within 90 days of the date your ticket was issued” by completing an online form.

Moffatt then sued for the fare difference, prompting Air Canada to issue what the tribunal member Christopher Rivers called a “remarkable submission” in its defense.

Air Canada argued that despite the error, the chatbot was a “separate legal entity” and thus was responsible for its actions.

While Air Canada argued correct information was available on its website, Rivers said the company did “not explain why the webpage titled ‘Bereavement Travel’ was inherently more trustworthy” than its chatbot.


The original article contains 414 words, the summary contains 164 words. Saved 60%. I'm a bot and I'm open source!

[–] [email protected] -1 points 10 months ago

They weren't misled. That implies that the bot was able to understand company policy.