News

Air Canada attempted to argue that it wasn’t accountable for the inaccurate assurances made by its chatbot — however, that defense didn’t succeed in court

 

Air Canada attempted to argue that it wasn’t accountable for the inaccurate assurances made by its chatbot — however, that defense didn’t succeed in court

 

If chatbots are set to supplant human customer service, they must ensure the accuracy of the information provided

Air Canada flight delays: Tech issue to blame | CTV News

Can AI chatbots ease the burden on customer service representatives? Lots of businesses seem to think so, but they better hope their chatbots don’t cause the same problem as Air Canada’s. The airline has just been forced to offer a partial refund to a customer, honoring a refund policy that its chatbot seemingly made up on the spot.

The incident in question happened to Jack Moffat, who went to Air Canada’s chatbot to help him understand the airline’s bereavement travel policy following the death of his grandmother. The chatbot explained that it was possible to book a flight immediately and request a partial refund within 90- days.

Unfortunately Air Canada’s bereavement travel policy states the airline won’t provide refunds for travel after the flight has been booked. Because of this Moffatt’s refund request was rejected. Air Canada admitted the bot was at fault, and instead offered a $200 flight voucher and promised to update the chatbot to ensure this doesn’t happen again.

Read Also:  

In just 15 minutes, this beginner kettlebell workout can help you develop a stronger physique

Air Canada’s key argument was that because the chatbot had linked to the official bereavement policy page, Moffatt should have known what the situation was. It also argued that the chatbot shouldn’t have been trusted in the first place, and that the airline could not be held responsible since the bit was a “separate legal entity”.

Obviously the tribunal wasn’t buying this excuse, especially since Air Canada didn’t explain why it’s not liable for information provided by its agents — including chatbots and human customer service representatives. The tribunal also ruled that Air Canada also didn’t explain why customers should be responsible for fact checking, and that Moffatt had no reason to believe the chatbot wasn’t supplying accurate information.

Moffatt was awarded a partial refund to the sum of $650.88 CAD ($428.29), plus additional damages to cover interest and the tribunal fees.

It’s rather bewildering that Air Canada would ever try to argue this in court. The idea that a company can offer a chatbot in place of a human customer service rep, then refuse to accept responsibility for when things go wrong, is pretty ridiculous. Especially when such a small amount of money is on the line.

Adam Leon Smith, Chair of F-TAG, the technical advisory group of BCS and a leading expert in AI safety, told Tom’s Guide that “It is amazing Air Canada even tried to fight this claim. In this context, advice given by a chatbot is obviously equivalent in standing to advice given on their web page.” Later adding that, “deployers of AI need to understand its limitations, and that they ultimately remain responsible for mistakes AI makes on their behalf”.

“It is amazing Air Canada even tried to fight this claim.

Adam Leon Smith, Chair of F-TAG
Meanwhile Ryan Carrier, CEO of AI safety and certification organization forHumanity, similarly criticized Air Canada’s response and the precedent it could have set if successful. Carrier told Tom’s Guide that forHumanity has “consistently argued that all tools such as AI, Algorithmic, and Autonomous Systems should always have a responsible, accountable beneficial owner”. In other words, companies need to know the risks of using chatbots and similar tools, and if your chatbot is going to go off the rails and make a bunch of promises you better be willing to back them up.

Or, you know, stop trying to cut corners in customer service and hire extra human operatives who can be trained and instructed to not make up company policy on the fly. That way customers know what the deal is and companies don’t have to deal with quite as many irate complaints over false promises.

Plus human operators can be fired if they swear at customers, or spend write poetry that criticizes their employer during work hours.

Related Articles

Leave a Reply

Your email address will not be published. Required fields are marked *

Back to top button