Air New Zealand may have made some of the most memorably wacky in-flight safety videos in the industry, but its first foray into chatbots, via 'Bravo Oscar Tango' – Oscar for short – doesn't seem quite such a success.
When questioned by this reporter (as seen in the screenshots above), Oscar was unable to answer basic questions about flight bookings unless they were asked in an exact scripted phrase. Oscar also couldn't recognise abbreviated spellings of city names, 'think ahead' to provide a few options at a time, or reuse information previously entered in the same chat, despite supposedly being a 'learning' bot.
Oscar is not the only bot in the market showing obvious flaws. But the limitations of this technology don't seem to be discouraging marketers from investing.
Chatbots have come a long way since Eliza, an application built at MIT in 1966 that could carry out a text dialogue with human interlocutors. They are now on the up and up, proliferating on messaging apps like Facebook Messenger, WeChat, Kik and Line, and predicted to have a transformational impact on the digital workplace in the next five to 10 years, according to Gartner's 'Hype Cycle', which looks at the bell curve of emerging technologies.
Spend a few minutes with your average chatbot, however, and you will quickly realise that they cannot sustain a real conversation. Yeong Yee, senior consultant at Ogilvy & Mather Singapore, wrote as much in an opinion column for Campaign earlier this year: "Chatbots are perennially question-and-answer-based. Ask a question or provide a trigger, and the chatbot gives an answer. Often, conversations do not extend beyond this scripted give-and-take."
But customers still want simplistic chatbots to 'respect' them, no matter how unpredictable their behaviour. Developing chatbots that somehow "know" how customers feel is crucial, said Gentle at a recent Shanghai event. Brands can do this by using language cues to 'sense' emotions, then respond accordingly. If a user expresses anger, for example, an emotionally-intelligent chatbot would try to resolve the situation through further questioning or pass the baton to a human customer service agent, he said.
Most chatbots do not elicit this level of understanding, leading to potential consumer disappointment. "The conversational manner of the user interface lulls us into imagining we are talking with 'someone'," says Yee. "We expect it to display empathy. Or maybe, regardless of how the chatbot is designed, we want to break them to test the limits of their ability to sustain a human facade."
What is a bot?
At its simplest, a chatbot is a computer programme that mimics conversations with people and is used as a means of automatically dispensing information. It works by picking up keywords that, when entered by users in a graphical user interface (GUI, pronounced 'gooey', also known as a human-computer interface), trigger automated replies.
All chatbots are bots, but not all bots are chatbots, clarifies Jeffrey Broer, founder of bot-development platform Recime.io. He identifies three types of bots: task bots, option bots and chatbots. While all three are good for aggregating data to qualify leads and optimise conversions, chatbots are "the new apps," says Broer, and conversation "the new interface".
Pharma firm Boehringer Ingelheim (BI), for example, developed a Facebook Messenger chatbot called Tabatha in May this year to allow people with asthma to engage with the brand in a more private environment, based on the insight that sufferers tend to accept asthma attacks as the norm, meaning many don't bother seeking treatment.
Previously, since the campaign first launched in 2014, the brand has been driving users to an interactive website where they could self-identify as symptomatic by responding to the Royal College of Physician’s ‘3 Questions’ (RCP3Qs).
Switching an online questionnaire for a conversation with a Facebook chatbot was a timely move, says Valerie Hargreaves, senior director of healthcare in Cohn & Wolfe (BI's agency), given both the increasing popularity of ‘dark social’ - private social channels where conversations are not visible to the wider public (900 million people a month use Facebook Messenger, according to eMarketer) - and people's apparent willingness to engage with AI about their health.
A chatbot can collect richer data and enable deeper measurement of campaign success. In BI's case, this included capturing consumers' intentions to act as a result of the information received for the first time, adds Hargreaves, who said Tabatha generated over double the number of engagements on World Asthma Day in 2017 compared to 2016, without Tabatha.
After privacy, proactiveness is another vital attribute of a successful healthcare chatbot. The Hong Kong branch of French insurance firm AXA, for example, created an 'intelligent' personal coachbot called Alex as part of the brand's wellness coaching app, which claims to seamlessly connect with over 150 wearable devices including Fitbit and Apple Health.
Apart from answering broad questions on wellness goals and objectives, Alex can apparently tackle more abstract issues relating to gym routines, healthy food recommendations or even what sporting activities are available in Hong Kong. It can even initiate suggestions to users based on observations of their behaviour recorded on other devices (such as last night's sleep pattern, missed schedules, or the time of day or week), according to Brady Ambler, strategy director with Publicis Worldwide Hong Kong, who worked on the campaign.
A final consideration that's "critical" for healthcare bots is the design of a really detailed response framework that can anticipate sensitive situations, says Ambler. This means ensuring the chatbot knows exactly when to transfer the conversation to a human colleague, a protocol many bots lack. It's part of the reason that once the novelty wears off, many users will regard chatbots as an "useless" brand interaction point, remarks Forrester Research senior analyst Xiaofeng Wang.
Training a bot is "vastly different from just content marketing on Facebook"
However, once chatbots are capable of inferring user context and intent in the near future, marketing use cases will grow in complexity and sophistication, thinks Wang. Think minimum viable product (MVP) and iterate as you learn, advises Ask. "Building a chatbot with a conversational interface is even less one-and-done than an app."
As chatbots evolve, the best will depend heavily both on AI and on the humans who teach the machine. Says Ambler: "I believe we’ll see our first super chatbot within the next 5 years. Something that makes the chatbots of today seem cute and antiquated."
CASE STUDY: Lost in translation in Hong Kong
Replicating the intricacies of human language is no small task, and having a clueless chatbot master the nuances of varied speech patterns, slang, sarcasm and insinuation common in Hong Kong, a city already fragmented with multiple marketing languages, is even more difficult.
Kinni Mew, CEO of NLP technology provider Mindlayer, says one of his projects almost got killed because the company committed some NLP mistakes in Cantonese. Speaking at a recent startup event, he compiles these rules for chatbot marketers in Hong Kong:
You'll fail if you think your dataset is ready.
Many brands, especially banks and insurers in Hong Kong, do not yet have usable data to train their chatbots. Audio archives of phone recordings with customers do not count as data, for example. "We rely on text to process syntax. Banks and insurers, both in highly-regulated industries, cannot use speech-to-text conversion," he says.
You'll fail if you think users will follow chatbot instructions.
Say a chatbot issues the instruction that a user should enquire only about policy details or promotion mechanics. Fine, but Hong Kong users will always ask more than the scope of a chatbot. "For one client, initially we typed everything using a FAQ format, but we discovered that a lot of customers are asking other questions outside of the agreed FAQ. There needs to be a lot of iteration. You have to release an updated chatbot version every single day so people can train the chatbot," adds Mew.
You'll fail if you treat Cantonese as English.
When Cantonese is translated into English, it becomes meaningless. "Garbage in, garbage out, we say. Brands can’t just purely translate user input. You need to have data from scratch to train your chatbot," Mew says.
You'll fail if you expect users to follow a defined language.
Hong Kong consumers usually type in broken sentences, mixing languages. Chatbots judge things in a binary fashion: it's either this or that, meaning untrained bots are unable to handle 'Hongklish' amalgamations, typos or emoticons. Probability analysis helps determine fuzzy logic for every single word or character, and deep learning helps make a word coherent enough to look beyond spelling errors, says Mew, but there is still some way to go.
You'll fail if your chatbot doesn't have a strong use case.
When your chatbot sucks, Hong Kongers will drop out immediately, so any sales are just a pipe dream.
You'll fail if you measure the wrong metrics.
"Wrong metrics" include things like asking customers to rate the conversation after a chatbot session. A plain numbered grade does not reveal the customer’s pain points, such as whether they like the chatting experience because they don’t have to call the bank's call centre, but maybe they still haven’t solved their core problem of a wrong account balance or a lost credit card. Brands need to have very specific questionnaires when they ask Hong Kongers for feedback.
You'll fail if you launch your chatbot before it is ready.
The term ‘ready’ means different things to different marketers in Hong Kong. To some, a chatbot may be "99 percent ready" based on accuracy testing with a 100-person sample. But this is not complete, Mew emphasises. A template of test scenarios should be built everyday and fed with fresh data to continously train the chatbot, he says. "The best tester is your actual audience, of course. So you have to strike a balance in terms of launch date."
You'll fail if you don't think 'small talk' is important.
Brands must anticipate ‘stupid' questions by users, like "who is your boss?" or "are you a boy or a girl?". Also, because people know they are not talking to a real human, they are cavalier about the consequences of improper etiquette. People may deliberately swear or say something inappropriate to the chatbot. "For this reason, web scraping (from Quora, Facebook, online forums, etc) to respond is very dangerous in case you scrape foul language or negative political opinions from these internet sites in order to handle these", warns Mew.