Professional Documents
Culture Documents
Himanshu Kumar Task 2
Himanshu Kumar Task 2
Contents
ABSTRACT...............................................................................................................................2
INTRODUCTION......................................................................................................................2
CHATBOTS BACKGROUND.................................................................................................3
METHODOLOGY.....................................................................................................................3
STAGE 1: GATHERING INFORMATION.........................................................................3
STAGE 2: ARTICLE FILTERING AND REVIEWING......................................................4
LITERATURE REVIEW ANALYSIS......................................................................................4
IMPLEMENTATION APPROACHES TO CHATBOT.......................................................5
RULE BASED CHATBOTS.................................................................................................5
ARTIFICIAL INTELLIGENCE CHATBOTS......................................................................5
DATASETS USED....................................................................................................................5
EVALUATION..........................................................................................................................6
DISCUSSION............................................................................................................................6
CONCLUSION..........................................................................................................................7
References..................................................................................................................................8
2
ABSTRACT
Chatbots are active conversational computer systems that are programmed to mimic human
speech and provide automatic online advice and assistance. Open data refers data that can be
freely used, re-used, and redistributed by anyone - subject only, at most, to the requirement to
attribute and share alike. Due to the general rising benefits of chatbots, numerous sectors
have adopted them to provide virtual assistance to customers. Chatbots use methods and
algorithms from two Artificial Intelligence domains: Machine Learning and Natural
Language Processing. However, there are numerous obstacles and limitations in their use. In
this survey, we will look at some of the most recent advancements in chatbots, including
Artificial Intelligence and Natural Language Processing. We will discuss the limitations and
benefits of integrating open data with chatbots and, the major highlights, issues and
limitations of existing research and offer suggestions for future research. (Keyner)
INTRODUCTION
A wide range of industries and applications are presently using chatbots, from education and
e-commerce to healthcare and entertainment. As a result, chatbots like as Mitsuku & Jessie
Humani, which are geared toward "small talk" and intended to foster a sense of human
engagement, can offer both support and fun to their users. It appears that chatbots are more
interesting to users than a website's conventional Constantly Asked Questions (FAQ) page.
Chatbots can assist numerous customers at once, making them more efficient and less
expensive than human customer service representatives. For the end user, chatbots can also
be utilised to provide amusement and friendship in addition to customer support and help. In
spite of this, chatbots appear to have varied levels of human-likeness, as well as different
levels of disclosure, which affects how and when users engage with and trust chatbots.
It has grown increasingly widespread in recent years because to the commoditization and
increase in computational power as well as the spreading of open-source tools and
frameworks. As AI and NLP approaches continue to improve, chatbots are becoming more
adaptable in terms of applicability and maintainability, while also becoming better at
mimicking human communication. A few areas for development include contextual &
emotional understanding & gender biases when it comes to human-chatbot interfacing.
Chatbots, on the other hand, lack the ability to recognise conversational context or emotional
linguistic cues, which hinders their ability to communicate in a nice and entertaining manner.
Chatbots, on the other hand, preferred to receive on traditionally feminine roles, which they
do with traditionally feminine traits and often demonstrate stereotypical behaviour, revealing
a bias against women in chatbots' implementation and use. (Følstad)
Improvements in the implementation and evaluation of chatbots are essential research
subjects since they are widely used and implemented in so many different fields. Deep
Learning algorithms are the focus of this paper's review of the literature on chatbots, which
includes an in-depth look at current implementation methods, as well as an examination of
the limitations and challenges of chatbot implementation and use, as well as suggestions for
future research into the topic. This paper also makes recommendations for future research
into chatbots.
3
CHATBOTS BACKGROUND
In 1950, after he presented the question "Can computers think?" Alan Turing is regarded to
have been the first individual to conceptualise the concept of a chatbot, although this pursuit
has profound origins in human history. It's easy to associate Turing's depiction of an
intelligent machine's behaviour with the widely accepted notion of a chatbot.
As computational power and Natural Language tools and methods have improved, chatbots
have evolved. In 1966, the creation of ELIZA resulted in the first successful deployment of a
chatbot that mainly depended on linguistic rules & pattern matching techniques. Using a
keyword-matching algorithm, it could connect with users. In order to deliver a response to the
user, it looks for an acceptable transformation rule. Eliza was a ground breaking system that
sparked a wave of innovation in the industry. Because it relied on limited context
identification and pattern matching criteria, ELIZA's reach of knowledge was constrained
because it couldn't be easily deployed in other fields. (Adam)
In the 1980s, the application of Artificial Intelligence (AI) in chatbots was a significant
development. Because A.L.I.C.E is an extension of XML, it uses the Artificial Intelligence
Scripting Language, or AIML for short. A.L.I.C.E.'s software could benefit from the addition
of dialogue pattern information thanks to this programme, which was created specifically for
that purpose. For data objects in AIML, themes and categories are used as the building
blocks. Essentially, categories are the building blocks of knowledge, that are made up of a
rule to connect user inputs to the chatbot's outputs. The user input is expressed by rule
pattern, while the chatbot's output is specified by rule templates, A.L.I.C.E knowledge base.
Because of the ease with which the AIML knowledgebase could be expanded to
accommodate new data objects, it represented a considerable improvement over earlier
pattern matching systems. In addition, more Loebner Prize-winning chatbots were built with
Chat Script, the successor to AIML.
Essentially, the goal behind this cutting-edge technology was to produce the desired output
by matching user-supplied text to a certain topic and applying rules relevant to that topic. In
the wake of Chat Script, chatbots entered a new age of technological advancement. Semantic
analysis and comprehension began to take precedence. (AbejideAde-Ibijola)
METHODOLOGY
There are two parts to the methodology we used to gather the data for this literature review.
Each stage has a variety of tasks. Search terms relevant to the subject matter and databases of
research publications are first identified. Then, we gather research articles about chatbots
from the databases we've chosen. The goal of these activities is to obtain as much information
as possible about the subject. Analysis of retrieved articles is the focus of our second phase of
research. According to the four features of chatbots that have been documented in literature:
design, implementation, application, and assessment methodologies, we categorise the papers
into four groups. We'll go into more depth about each of these tasks in the sections that
follow.
STAGE 1: GATHERING INFORMATION
To locate studies on chatbots, we combed through the databases of three major publishers.
ScienceDirect and Springer are two of the most popular. Peer-reviewed works in the areas of
Linguistics, Artificial Intelligence, & Human-Computer Interaction can be found in these
4
Finally, we looked through the bibliography of the 115 papers to see if there were any other
relevant sources. On Google Scholar, we found full-text publications that appeared in the
bibliographies of 115 papers, and we retrieved them. We were able to find an additional 201
related articles through this technique. 316 articles in all were available for evaluation,
making this a massive undertaking.
5
DATASETS USED
Datasets that appear to be the most often utilised to train deep learning chatbots models will
be discussed in this section. To begin, we must distinguish between datasets that are open to
the public and those that are private. There seem to be just few open domain databases that
are often utilised in the literature:
Daily-Dialog, Open-Subtitles, and Cornell are all part of this project. We'll go into greater
detail about each of these datasets. Contrary to popular belief, no closed domain datasets
were found to be frequently referenced in the research reviewed. In some cases, this is
because the datasets in question are so narrowly focused on a single issue or goal. For the
sake of time, we won't go into great detail about any of these datasets.
Open-Subtitles. In a number of works of literature, the Open-Subtitles collection is used.
XML-formatted movie subtitles are stored in an open-source database. Many languages' open
access movie subtitles can be found in this database. There appear to be millions of sentence
pairs in the dataset, but the data appears to be of poor quality; scene descriptions, nearer
captioning, & segmented sentences appear in the conversations, which can be troublesome
when trying to train an open domain chatbot, because the coherency of the dialogue can be
lost. This is a problem when trying to train an open domain chatbot.
EVALUATION
Since human communication serves a variety of purposes, evaluating dialogue technologies
has proven to be difficult. Metrics for evaluating a chatbot's performance can vary based on
its intended purpose. The effectiveness of a chatbot's engagement with a user will be a major
factor in determining its success. In contrast, a companion chatbot would be judged on its
own ability to keep the discussion continuing and to keep people engaged. Human evaluation
& automated evaluation metrics are the two primary methods for evaluating a chatbot.
DISCUSSION
The architecture of chatbots still has some flaws, despite recent advances in Deep Learning &
Natural Language Processing. Because of an improper approach to dialogue modelling,
different language models offered as the architecture of chatbots are still unable to accurately
simulate human interaction. For example, this model attempts to tackle conversational
difficulties by using a "next-step" method in which it aims to predict the greatest possible
answer given an input. As a rule of thumb, however, this is not how human conversation
works, which takes into account a succession of preceding stages, the context in which it is
taking place, as well as the information being exchanged among the participants. Chatbot
designs portray human conversation as a series of discrete steps, but in reality, human
interaction is more like a continuous back-and-forth, with each step dependent on the one
before it or the one after it. Even while some new versions have attempted to capture this
conversational context, there is still a long way to go.
The apparent absence of a learnt AI model in Information Retrieval chatbots is another
critical flaw in the architecture of chatbots. Because they can provide coherent answers to
specific topics (e.g., booking an appointment or product specifics, returning an item, &
finding learning materials), these chatbots are highly popular all over industries (e.g.
healthcare, e-commerce, customer support and education) because they can find a similar
7
CONCLUSION
We've done a literature review and a state-of-the-art analysis of machine translation,
applications, datasets used, and assessment systems in this research. Moreover, we've
highlighted the present issues and limitations, and also highlighted gaps in the research.
Despite advances in AI, chatbots are unable to mimic human speech, despite advances in AI.
Dialog modelling has been done incorrectly, and there are no domain-specific data sets
available for free. Also, there is a lack of a learned AI model for Information Retrieval
chatbots. This kind of model could be put to use in a number of different fields. Despite
recent advances in the industry, there is still a chasm to be bridged in terms of application. It
takes a lot of computer power and training data to build large models. There is no standard
way to evaluate chatbots. However, human evaluation is costly, time consuming, difficult to
handle up, biased and lacks coherence for several models that rely on human evaluations.
There should be a new, dependable automated evaluation approach offered to overcome these
limitations. There is also a lack of data on most recent breakthroughs in language models that
could be applied to chatbot like Transformers, as indicated by recent studies. As a result, data
utilised to train multiple models must be thoroughly examined and analysed. Using this
method, researchers are able to more accurately compare the outcomes of various models. In
fact, it's hard to tell the difference between a chatbot's applications and a chatbot's social or
companion functions. It is an interesting challenge that combines Deep Learning & Natural
Language Processing to create chatbots. Chatbots have been around for sixty years, yet
they're still evolving and posing fresh challenges. Smaller, more flexible, and less domain-
8
dependent models could help bridge these gaps. Language models for industry-specific
applications, better human-like model topologies and better evaluation frameworks would be
significant advancements in the field of artificial intelligence.
References
AbejideAde-Ibijola, Chinedu WilfredOkonkwo. Chatbots applications in education: A
systematic review. 2 March 2021. Webpage. 11 November 2021.
Adam, Martin. AI-based chatbots in customer service and their effects on user compliance.
17 March 2020. https://link.springer.com/article/10.1007/s12525-020-00414-7. 21
May 2021.
Deloitte. Chatbots Point of View. 11 March 2018. PDF. 21 May 2020.
Følstad, Asbjørn. Future directions for chatbot research: an interdisciplinary research
agenda. 19 October 2019. PDF. 19 November 2021.
Keyner, Sophia. Open Data Chatbot. 16 May 2020. Document. 17 October 2019.
Xueming Luo, Siliang Tong, Zheng Fang, and Zhe Qu. Machines versus Humans: The
Impact of AI Chatbot Disclosure on Customer Purchases. 6 JUNE 2019. PDF. 11
MARCH 2021.