In so many ways I think chatbots are only just getting started – their potential is much underestimated at present. A big challenge is for chatbots mature so that they do more than is possible as a result of content entry wizards. If your content is created with a few easy clicks, it is unlikely to be much inspiration to anyone – and to date, despite much work in the field, the ability to emulated the creative open ended nature of real intellingence has seen only very partial success.
Indeed, this is one of the key benefits of chatbots – providing a 24/7/365 presence that can give prospects and customers access to information no matter when they need it. This, in turn, can result in cost-savings for companies that deploy chatbots, as they cut down on the labour-hours that would be required for staff to manage a direct messaging service every hour of the week.
Companies most likely to be supporting bots operate in the health, communications and banking industries, with informational bots garnering the majority of attention. However, challenges still abound, even among bot supporters, with lack of skilled talent to develop and work with bots cited as a challenge in implementing solutions, followed by deployment and acquisition costs, as well as data privacy and security.
Furthermore, major banks today are facing increasing pressure to remain competitive as challenger banks and fintech startups crowd the industry. As a result, these banks should consider implementing chatbots wherever human employees are performing basic and time-consuming tasks. This would cut down on salary and benefit costs, improve back-office efficiency, and deliver better customer care.
From any point in the conversation, the bot needs to know where to go next. If a user writes, “I’m looking for new pants,” the bot might ask, “For a man or woman?” The user may type, “For a woman.” Does the bot then ask about size, style, brand, or color? What if one of those modifiers was already specified in the query? The possibilities are endless, and every one of them has to be mapped with rules.
One key reason: The technology that powers bots, artificial intelligence software, is improving dramatically, thanks to heightened interest from key Silicon Valley powers like Facebook and Google. That AI enables computers to process language — and actually converse with humans — in ways they never could before. It came about from unprecedented advancements in software (Google’s Go-beating program, for example) and hardware capabilities.
“The chat space is sort of the last unpolluted space [on your phone],” said Sam Mandel, who works at the startup studio Betaworks and is also building a weather bot for Slack called Poncho. “It’s like the National Park of people’s online experience. Right now, the way people use chat services, it’s really a good private space that you control.” (That, of course, could quickly go sour if early implementations are too spammy or useless.)
No one wants to download another restaurant app and put in their credit-card information just to order. Livingston sees an opportunity in being able to come into a restaurant, scan a code, and have the restaurant bot appear in the chat. And instead of typing out all the food a person wants, the person should be able to, for example, easily order the same thing as last time and charge it to the same card.
I will not go into the details of extracting each feature value here. It can be referred from the documentation of rasa-core link that I provided above. So, assuming we extracted all the required feature values from the sample conversations in the required format, we can then train an AI model like LSTM followed by softmax to predict the next_action. Referring to the above figure, this is what the ‘dialogue management’ component does. Why LSTM is more appropriate? — As mentioned above, we want our model to be context aware and look back into the conversational history to predict the next_action. This is akin to a time-series model (pls see my other LSTM-Time series article) and hence can be best captured in the memory state of the LSTM model. The amount of conversational history we want to look back can be a configurable hyper-parameter to the model.
Feine, J., Morana, S., and Maedche, A. (2019). “Leveraging Machine-Executable Descriptive Knowledge in Design Science Research ‐ The Case of Designing Socially-Adaptive Chatbots”. In: Extending the Boundaries of Design Science Theory and Practice. Ed. by B. Tulu, S. Djamasbi, G. Leroy. Cham: Springer International Publishing, pp. 76–91. Download Publication
A chatbot that functions through machine learning has an artificial neural network inspired by the neural nodes of the human brain. The bot is programmed to self-learn as it is introduced to new dialogues and words. In effect, as a chatbot receives new voice or textual dialogues, the number of inquiries that it can reply and the accuracy of each response it gives increases. Facebook has a machine learning chatbot that creates a platform for companies to interact with their consumers through the Facebook Messenger application. Using the Messenger bot, users can buy shoes from Spring, order a ride from Uber, and have election conversations with the New York Times which used the Messenger bot to cover the 2016 presidential election between Hilary Clinton and Donald Trump. If a user asked the New York Times through his/her app a question like “What’s new today?” or “What do the polls say?” the bot would reply to the request.
Keep it conversational: Chatbots help make it easy for users to find the information they need. Users can ask questions in a conversational way, and the chatbots can help them refine their searches through their responses and follow-up questions. Having had substantial experience with personal assistants on their smartphones and elsewhere, users today expect this level of informal interaction. When chatbot users are happy, the organizations employing the chatbots benefit.
Businesses are no exception to this rule. As more and more users now expect and prefer chat as a primary mode of communication, we’ll begin to see more and more businesses leveraging conversational AI to achieve business goals—just as Gartner predicts. It’s not just for the customer; your business can reduce operational costs and scale operations as well.
It’s not all doom and gloom for chatbots. Chatbots are a stopgap until virtual assistants are able to tackle all of our questions and concerns, regardless of the site or platform. Virtual assistants will eventually connect to everything in your digital life, from websites to IoT-enabled devices. Rather than going through different websites and speaking to various different chatbots, the virtual assistant will be the platform for finding the answers you need. If these assistants are doing such a good job, why would you even bother to use a branded chatbot? Realistically this won’t take place for sometime, due to the fragmentation of the marketplace.
Just last month, Google launched its latest Google Assistant. To help readers get a better glimpse of the redesign, Google’s Scott Huffman explained: “Since the Assistant can do so many things, we’re introducing a new way to talk about them. We’re them Actions. Actions include features built by Google—like directions on Google Maps—and those that come from developers, publishers, and other third parties, like working out with Fitbit Coach.”
Developed to assist Nigerian students preparing for their secondary school exam, the University Tertiary Matriculation Examination (UTME), SimbiBot is a chatbot that uses past exam questions to help students prepare for a variety of subjects. It offers multiple choice quizzes to help students test their knowledge, shows them where they went wrong, and even offers tips and advice based on how well the student is progressing.
One pertinent field of AI research is natural language processing. Usually, weak AI fields employ specialized software or programming languages created specifically for the narrow function required. For example, A.L.I.C.E. uses a markup language called AIML, which is specific to its function as a conversational agent, and has since been adopted by various other developers of, so called, Alicebots. Nevertheless, A.L.I.C.E. is still purely based on pattern matching techniques without any reasoning capabilities, the same technique ELIZA was using back in 1966. This is not strong AI, which would require sapience and logical reasoning abilities.