Smart chatbots rely on artificial intelligence when they communicate with users. Instead of pre-prepared answers, the robot responds with adequate suggestions on the topic. In addition, all the words said by the customers are recorded for later processing. However, the Forrester report “The State of Chatbots” points out that artificial intelligence is not a magic and is not yet ready to produce marvelous experiences for users on its own. On the contrary, it requires a huge work:
Simplified and scripted. Chatbot technology is being tacked on to the broader AI message, and while it’s important to note that machine learning will help chatbots get better at understand and responding to questions, it’s not going to make them the conversationalists we dream them to be. No matter what the marketing says, chatbots are entirely scripted. User says x, chatbot responds y.
The front-end app you develop will interact with an AI application. That AI application—usually a hosted service—is the component that interprets user data, directs the flow of the conversation and gathers the information needed for responses. You can then implement the business logic and any other components needed to enable conversations and deliver results.
[In] artificial intelligence ... machines are made to behave in wondrous ways, often sufficient to dazzle even the most experienced observer. But once a particular program is unmasked, once its inner workings are explained ... its magic crumbles away; it stands revealed as a mere collection of procedures ... The observer says to himself "I could have written that". With that thought he moves the program in question from the shelf marked "intelligent", to that reserved for curios ... The object of this paper is to cause just such a re-evaluation of the program about to be "explained". Few programs ever needed it more.
The biggest benefit of having a conversational AI solution is the instant response rate. Answering queries within an hour translates into 7X increase in the likelihood of converting a lead. Customers are more likely to talk about a negative experience than a positive one. So nipping a negative review right in the bud is going to help improve your product’s brand standing.
Unfortunately, my mom can’t really engage in meaningful conversations anymore, but many people suffering with dementia retain much of their conversational abilities as their illness progresses. However, the shame and frustration that many dementia sufferers experience often make routine, everyday talks with even close family members challenging. That’s why Russian technology company Endurance developed its companion chatbot.
L’usage des chatbots fut d’abord en partie expérimental car il présentait un certain risque pour les marques en fonction des dérapages sémantiques possibles et des manipulations ou détournements également envisageables de la part des internautes. Les progrès dans le domaine ont cependant été rapides et les chatbots s’imposent désormais dans certains contextes comme un nouveau canal de support ou contact client garantissant disponibilité et gains de productivité.
Context: When a NLU algorithm analyzes a sentence, it does not have the history of the user conversation. It means that if it receives the answer to a question it has just asked, it will not remember the question. For differentiating the phases during the chat conversation, it’s state should be stored. It can either be flags like “Ordering Pizza” or parameters like “Restaurant: ‘Dominos’”. With context, you can easily relate intents with no need to know what was the previous question.
ELIZA's key method of operation (copied by chatbot designers ever since) involves the recognition of clue words or phrases in the input, and the output of corresponding pre-prepared or pre-programmed responses that can move the conversation forward in an apparently meaningful way (e.g. by responding to any input that contains the word 'MOTHER' with 'TELL ME MORE ABOUT YOUR FAMILY'). Thus an illusion of understanding is generated, even though the processing involved has been merely superficial. ELIZA showed that such an illusion is surprisingly easy to generate, because human judges are so ready to give the benefit of the doubt when conversational responses are capable of being interpreted as "intelligent".