L’usage des chatbots fut d’abord en partie expérimental car il présentait un certain risque pour les marques en fonction des dérapages sémantiques possibles et des manipulations ou détournements également envisageables de la part des internautes. Les progrès dans le domaine ont cependant été rapides et les chatbots s’imposent désormais dans certains contextes comme un nouveau canal de support ou contact client garantissant disponibilité et gains de productivité.
Artificial Intelligence is currently being deployed in customer service to both augment and replace human agents - with the primary goals of improving the customer experience and reducing human customer service costs. While the technology is not yet able to perform all the tasks a human customer service representative could, many consumer requests are very simple ask that sometimes be handled by current AI technologies without human input.
Companies and customers can benefit from internet bots. Internet bots are allowing customers to communicate with companies without having to communicate with a person. KLM Royal Dutch Airlines has produced a chatbot that allows customers to receive boarding passes, check in reminders, and other information that is needed for a flight. Companies have made chatbots that can benefit customers. Customer engagement has grown since these chatbots have been developed.
Modern chatbots are frequently used in situations in which simple interactions with only a limited range of responses are needed. This can include customer service and marketing applications, where the chatbots can provide answers to questions on topics such as products, services or company policies. If a customer's questions exceed the abilities of the chatbot, that customer is usually escalated to a human operator.
Of course, it is not so simple to create an interactive agent that the user will really trust. That’s why IM bots have not replaced all the couriers, doctors and the author of these lines. In this article, instead of talking about the future of chatbots, we will give you a short excursion into the topic of chatbots, how they work, how they can be employed and how difficult it is to create one yourself.
To envision the future of chatbots/virtual assistants, we need to take a quick trip down memory lane. Remember Clippy? Love him or hate him, he’s ingrained in our memory as the little assistant who couldn’t (sorry, Clippy.). But someday, this paper clip could be the chosen one. Imagine with me if you will a support agent speaking with a customer over the phone, or even chat support. Clippy could be listening in, reviewing the questions the customer is posing, and proactively providing relevant content to the support agent. Instead of digging around from system to system, good ‘ole Clippy would have their back, saving them the trouble of hunting down relevant information needed for the task at hand.
The trained neural network is less code than an comparable algorithm but it requires a potentially large matrix of “weights”. In a relatively small sample, where the training sentences have 150 unique words and 30 classes this would be a matrix of 150x30. Imagine multiplying a matrix of this size 100,000 times to establish a sufficiently low error rate. This is where processing speed comes in.
Like apps and websites, bots have a UI, but it is made up of dialogs, rather than screens. Dialogs help preserve your place within a conversation, prompt users when needed, and execute input validation. They are useful for managing multi-turn conversations and simple "forms-based" collections of information to accomplish activities such as booking a flight.
The bottom line is that chatbots have completely transformed the way companies interact with their consumers. And guess what? This is just the very beginning. And the truth is that even though to some company leaders it may seem challenging to incorporate the omnichannel customer experience, it opens up a fantastic opportunity that allows businesses to engage with customers in a fresh, modern way. The outcome of this may prove to be an excellent opportunity to build more meaningful and long-lasting relationships with the customers.
You can structure these modules to flow in any way you like, ranging from free form to sequential. The Bot Framework SDK provides several libraries that allows you to construct any conversational flow your bot needs. For example, the prompts library allows you to ask users for input, the waterfall library allows you to define a sequence of question/answer pair, the dialog control library allows you to modularized your conversational flow logic, etc. All of these libraries are tied together through a dialogs object. Let's take a closer look at how modules are implemented as dialogs to design and manage conversation flows and see how that flow is similar to the traditional application flow.
Students from different backgrounds can share their views and perspectives on a specific matter while a chatbot can still adapt to each one of them individually. Chatbots can improve engagement among students and encourage interaction with the rest of the class by assigning group work and projects - similarly to what teachers usually do in regular classes.
In a traditional application, the user interface (UI) is a series of screens. A single app or website can use one or more screens as needed to exchange information with the user. Most applications start with a main screen where users initially land and provide navigation that leads to other screens for various functions like starting a new order, browsing products, or looking for help.
Feine, J., Morana, S., and Maedche, A. (2019). “Leveraging Machine-Executable Descriptive Knowledge in Design Science Research ‐ The Case of Designing Socially-Adaptive Chatbots”. In: Extending the Boundaries of Design Science Theory and Practice. Ed. by B. Tulu, S. Djamasbi, G. Leroy. Cham: Springer International Publishing, pp. 76–91. Download Publication
For each kind of question, a unique pattern must be available in the database to provide a suitable response. With lots of combination on patterns, it creates a hierarchical structure. We use algorithms to reduce the classifiers and generate the more manageable structure. Computer scientists call it a “Reductionist” approach- in order to give a simplified solution, it reduces the problem.
In the early 90’s, the Turing test, which allows determining the possibility of thinking by computers, was developed. It consists in the following. A person talks to both the person and the computer. The goal is to find out who his interlocutor is — a person or a machine. This test is carried out in our days and many conversational programs have coped with it successfully.
“Utility gets something done following a prompt. At a higher level the more entertainment-related chatbots are able to answer all questions and get things done. Siri and Cortana you can have small talk with, as well as getting things done, so they are much harder to build. They took years and years of giant company’s efforts. Different companies that don’t have those resources, like Facebook, will build more constrained utility bots.”
Typically, companies applied a passive engagement method with consumers. In other words, customer support only responds to complaining consumers – but never initiate any conversations or look for feedback. While this method was fine for a long while, it doesn’t work anymore with millennials. Users want to communicate with attentive brands who have a 24/7 support system and they won’t settle for anything less.
In one particularly striking example of how this rather limited bot has made a major impact, U-Report sent a poll to users in Liberia about whether teachers were coercing students into sex in exchange for better grades. Approximately 86% of the 13,000 Liberian children U-Report polled responded that their teachers were engaged in this despicable practice, which resulted in a collaborative project between UNICEF and Liberia’s Minister of Education to put an end to it.
However, as irresistible as this story was to news outlets, Facebook’s engineers didn’t pull the plug on the experiment out of fear the bots were somehow secretly colluding to usurp their meatbag overlords and usher in a new age of machine dominance. They ended the experiment due to the fact that, once the bots had deviated far enough from acceptable English language parameters, the data gleaned by the conversational aspects of the test was of limited value.