Professional Documents
Culture Documents
Victor Hugo C. de Albuquerque, Pethuru Raj, Satya Prakash Yadav (Eds.) Toward Artificial General Intelligence
Victor Hugo C. de Albuquerque, Pethuru Raj, Satya Prakash Yadav (Eds.) Toward Artificial General Intelligence
)
Toward Artificial General Intelligence
Also of interest
Artificial Intelligence for Virtual Reality
Edited by: Jude Hemanth, Madhulika Bhatia and Isabel De La Torre Diez,
ISBN ----, e-ISBN ----
Edited by
Victor Hugo C. de Albuquerque, Pethuru Raj,
Satya Prakash Yadav
Editors
Prof. Victor Hugo C. de Albuquerque
Department of Teleinformatics Engineering
Federal University of Ceará
Av. da Universidade 2853
60020-180 Fortaleza, Brazil
victor120585@yahoo.com.br
ISBN 978-3-11-132356-5
e-ISBN (PDF) 978-3-11-132374-9
e-ISBN (EPUB) 978-3-11-132416-6
www.degruyter.com
Preface
Artificial intelligence (AI) has been the prominent field of study and research in educa-
tional institutions and research labs across the globe. Strategically speaking, AI has the
wherewithal to bring a series of delectable and deft advancements across industry verti-
cals. As widely accepted, AI has turned out to be the prime and pioneering enabler of
business agility and automation. In other words, the real business transformation is to
be achieved through the smart leverage of the distinct AI power. AI is all about swift and
sagacious transitioning all kinds of intertwined digital data into timely and actionable
insights. The knowledge discovered through the growing array of AI technologies and
tools gets disseminated into all kinds of digital assistants, machineries at the manufactur-
ing floors, defense equipment, medical instruments, handhelds, wearables, portables,
mobile devices, robots, vehicles, drones, etc. Further on, business workloads and IT serv-
ices are empowered through timely and actionable insights to make them competent,
cognizant, and cognitive. AI intrinsically simplifies and speeds up the complicated pro-
cess of fulfilling the mantra of data-driven insights and insights-driven decisions. In
short, artistically replicating the contextually and comprehensively learning, articulat-
ing, and decision-making capabilities of the human brain in IT products, solutions, and
services is the principal role and responsibility of the AI paradigm. AI is becoming pene-
trative, pervasive, and persuasive too. All kinds of digital systems (the digitized version
of physical, mechanical, electrical, and electronics systems) are being empowered with
AI phenomenon to exhibit adaptive and autonomous competency in their everyday op-
erations, offerings, and outputs.
https://doi.org/10.1515/9783111323749-202
Contents
Preface V
List of authors IX
Index 401
List of authors
Dr. Abhisek Sethy Dr. Balambigai Subramanian
Department of CSE, Srinivasa Ramanujan Department of ECE, Kongu Engineering College,
Institute of Technology, Anantapur, Perundurai, Tamil Nadu, India
Andhra Pradesh, India sbalambigai@gmail.com
abhisek.sethy@silicon.ac.in
Dr. Balamurugan Easwaran
Dr. S. Ambareesh Texila American University, Zambia
School of Computer Science and Engineering, balamurugan.e@tau.edu.zm
JAIN (Deemed-to-be University), Bangalore,
Karnataka, India Dr. Balkeshwar Singh
ambihce@gmail.com Department of Mechanical Engineering,
Program of Manufacturing Engineering, Adama
Ambigaipriya S. Science and Technology University, Adama City,
Department of EEE, Mookambigai College of Ethiopia
Engineering, Pudukkottai, Tamil Nadu, India balkeshwar71@rediffmail.com
priyaeeemce@gmail.com
Bhola Khan
Dr. Amit Mahajan
Department of Economics, Yobe State University,
Centre for IT enabled services, University of
Damaturu, Nigeria
Jammu, Jammu Tawi, Jammu and Kashmir, India
bholakhan.apj@gmail.com
amit.mahajans@gmail.com
Dr. Bhoopathy Varadharajan
Dr. S. P. Anandaraj
Department of Computer Science and
Presidency University, Bangalore, Karnataka,
Engineering, Malla Reddy College of
India
Engineering, Secunderabad, Telangana, India
anandsofttech@gmail.com
v.bhoopathy@gmail.com
Dr. K. Arthi
Data Science and Business Systems, School of Dr. N. Bindu Madhavi
Computing, SRM Institute of Science and KL Business School and Programme Coordinator
Technology, Kattankulathur, Chennai, (MBA), KL Centre for Distance and Online
Tamil Nadu, India Education (CDOE), Koneru Lakshmaiah Education
arthik1@srmist.edu.in Foundation (Deemed to be University),
Andhra Pradesh, India
A. R. Arunarani dr.bindumadhavi@kluniversity.in
Department of Computational Intelligence,
School of Computing, College of Engineering Dr. S. Chandra Sekaran
and Technology, SRM Institute of Science and Department of Computer Science and
Technology, Kattankulathur campus, Chennai, Engineering, PSV College of Engineering and
Tamil Nadu, India Technology, Krishnagiri, Tamil Nadu, India
arunaraa@srmist.edu.in chandrudpi@gmail.com
https://doi.org/10.1515/9783111323749-204
X List of authors
Dr. Pethuru Raj is the chief architect in the Edge AI division of Reliance Jio
Platforms Ltd., Bangalore, India. He focuses on emerging technologies such as
the Internet of things (IoT), artificial intelligence (AI) model optimization, big
and streaming data analytics, blockchain, digital twins, cloud native, serverless
and edge computing, reliability engineering, microservices architecture (MSA),
event-driven architecture (EDA), 5G/6G, and metaverse. He previously worked
at IBM global Cloud Center of Excellence (CoE), Wipro Consulting Services
(WCS), and Robert Bosch Corporate Research (CR). He has published over 30
https://doi.org/10.1515/9783111323749-205
XIV About the editors
research papers in peer-reviewed journals, authored and edited 42 books, and contributed 54 book
chapters. He is a professional member of Association of Computing Machinery (ACM). He received his PhD
from Anna University, Chennai, India, and did his postdoctoral research in Computer Science and
Automation (CSA) at the Indian Institute of Science, Bangalore, and at the Nagoya Institute of Technology,
Kyoto University and at the University of Tsukuba, Japan.
https://doi.org/10.1515/9783111323749-001
2 Mahesh Maurya et al.
and present options. Cognitive computing then developed in the 1990s to give ma-
chines the ability to self-learn by recognizing patterns and adjusting behaviors. In the
2010s, ANNs – which mimic the neurons and synapses of the human brain – emerged
as the most sophisticated form of AI [3]. Deep learning has enabled machines to
achieve superhuman levels of accuracy in various tasks, including speech and image
recognition, language translation, and game playing (e.g., AlphaGo).
1.4 Taxonomy of AI
AI taxonomy is a system for categorizing AI technology. ML, knowledge representa-
tion and reasoning, NLP, computer vision, robotics, and human-computer interaction
1 Introduction to artificial intelligence 3
(HCI) are some of the categories into which it is typically separated. Then, in order to
provide a more thorough classification, each of these categories can be further broken
down into subcategories. For instance, supervised learning, unsupervised learning,
and reinforcement learning are three categories of ML [5]. Similarly, knowledge re-
presentation and reasoning can be divided into rule-based systems, expert systems,
ontologies, and logic [6]. By classifying AI technologies in a taxonomic manner, it be-
comes easier to compare and contrast different AI technologies, identify which tech-
nologies work better for which tasks, and identify areas for improvement which is
shown in Tables 1.1–1.5.
Classical AI Modern AI
Classical AI is a type of AI based on logic and Modern AI is a type of AI that uses more varied and
reasoning. It uses sets of rules or algorithms to complex methods to solve problems. It includes
solve problems in the same way a person with models such as deep learning, neural networks,
expertise in the topic might do. symbolic programming, genetic algorithms, and
machine learning.
Classical AI works by breaking down a problem and The solutions it learns and adjusts can be based on
using data to attempt to formulate a response. changing data sets and feedback from previous
tasks. This sort of AI is more independent and
adaptive than classical AI.
It often uses trial-and-error reasoning to figure out It is better equipped to handle difficult tasks like
the best solution. This type of AI is useful for image identification, self-driving cars, and even
sorting, searching, language processing, and other natural language processing. Modern artificial
tasks that require specific heuristics. intelligence is quickly taking over sectors like
healthcare, finance, and e-commerce.
Narrow AI General AI
Weak AI, sometimes referred to as narrow AI, Strong AI, sometimes referred to as general AI, is a
refers to artificial intelligence systems that have type of artificial intelligence that can comprehend
been created to do a very specific task without the and reason about a variety of activities.
assistance of a person.
These AI systems use a limited set of data to solve General AI has more advanced learning capabilities
a particular problem. than Narrow AI, and is able to improve its
performance based on progressive feedback.
Examples of narrow AI include machine learning Some examples of general AI systems include facial
(ML) algorithms, automated computer systems, recognition, natural language processing, virtual
and robots used to assemble a car or carry out assistants, and self-driving cars.
an industrial process.
4 Mahesh Maurya et al.
Weak AI Strong AI
Weak AI, sometimes referred to as narrow AI, is a A type of artificial intelligence called strong AI,
type of artificial intelligence that is concentrated on commonly referred to as artificial general
carrying out particular tasks autonomously. intelligence (AGI), is able to learn, reason, plan, and
solve problems in a way that is similar to how
humans do. It is regarded as artificial intelligence
technology‘s ultimate objective.
Weak AI systems are limited in their scope and are Strong AI requires advanced algorithms and
not capable of learning and adapting. computational power, and is still in its infancy in
terms of development.
Examples of weak AI include image recognition, Examples of strong AI include self-driving cars,
virtual personal assistants, medical diagnosis, and intelligent robots, and computer vision.
facial recognition technology.
Symbolic AI Sub-symbolic AI
Artificial intelligence that places a major emphasis Sub-symbolic AI, commonly referred to as “weak”
on the interpretation and manipulation of symbols or “bad” AI, is a classification of rtificial intelligence
is referred to as symbolic AI, also known as built on probabilistic algorithms and statistical
“strong” or “good” AI. models.
Narrow AI General AI
A branch of artificial intelligence known as machine Rule-based artificial intelligence (AI) is a more
learning (ML) gives computers the ability to learn traditional approach to AI and relies on a
from data, recognize patterns, and make choices on predetermined set of rules to allow computers to
their own. make decisions.
1 Introduction to artificial intelligence 5
Narrow AI General AI
It is based on the idea that a computer can learn Rule-based AI uses predefined sets of rules in
from data and automatically create computer order to generate an expected outcome, based on
algorithms to identify patterns and predict the available input data. This approach requires a
outcomes. lot of hand-coded rules and is less flexible than ML
algorithms that can adapt to changing conditions
by adapting to new data.
ML is a branch of AI that focuses on creating methods and algorithms that let com-
puters understand and use data to learn. The algorithms provide computers the abil-
ity to make decisions without explicit programming by allowing them to learn from
the facts at their disposal and from their prior experiences. ML methods include, for
instance, decision trees, neural networks, support vector machines, linear regression,
logistic regression, and k-means clustering. The use of neural networks, on the other
hand, is the main component of deep learning, a subset of ML that enables computers
to learn from complex data [7]. Deep learning algorithms use a variety of methods,
including hierarchical feature extraction and representation learning to help ma-
chines learn considerably more effectively. Convolutional neural networks, recurrent
neural networks, and generative adversarial networks are a few examples of deep
learning methods.
6 Mahesh Maurya et al.
The AI technology known as NLP makes it easier for computers to comprehend and
manipulate human language. This technology is widely employed in many different ap-
plications, including text summarization, speech recognition, machine translation, doc-
ument classification, and question answering. In order to identify correlations between
words and phrases, NLP technology analyzes the grammar, syntax, semantics, and
structure of written or spoken language. NLP has many uses, especially in industries
like healthcare, business, and education. In healthcare, for example, NLP is used for
clinical decision-making, medical information retrieval, research, and personalized
healthcare advice [8]. In business, NLP can be used to analyze customer comments and
questions, generate customer feedback, analyze customer sentiment, recommend prod-
ucts, and identify sales trends. In education, NLP can be used to help automate assess-
ment, improve search accuracy in libraries, and detect plagiarism in student work.
Natural language understanding (NLU) and natural language generation (NLG) are two
more technologies that are frequently used in conjunction with NLP. NLU assists in de-
coding the same input to determine the user’s intent and meaning, whereas NLG assists
in creating natural-sounding words and sentences from the data input [9]. Together,
these technologies allow computers to generate more accurate and meaningful results
from their input. NLP is a powerful tool that can be used to process and understand
natural language in a more efficient and accurate way. As more applications for NLP
are developed, its applications are likely to expand to a variety of industries. This tech-
nology can be used to create smarter and more efficient applications that deliver more
personalized experiences for users.
Computer vision and image recognition have become increasingly significant technol-
ogies in today’s society. Computer vision involves using software to recognize objects
within images and videos that can be further processed by a computer. It is a method
of “teaching” a computer to distinguish different objects within an image, such as a
bike, a car, or a person. Image recognition helps computers to determine character-
istics of a scene or object without any human input. By recognizing certain objects,
robots can then complete related tasks without the need for human help. Recently,
computer vision and image recognition technologies have seen a significant uptake
within the automotive and medical industries due to their high precision and accu-
racy in detecting and understanding objects. In cars, computer vision and image rec-
ognition can be used for automated emergency braking, active cruise control, parking
assistance, lane change assistance, and driver monitoring. In medicine, these technol-
ogies can be used to help quickly detect cancerous cells, precancerous cells, and any
abnormalities in X-rays or MRIs. Besides, computer vision and image recognition tech-
1 Introduction to artificial intelligence 7
nologies are also widely used for facial recognition, security, augmented reality (AR)/
virtual reality (VR) applications, robotics control, autonomous drones, deep learning
networks, and NLP. For example, facial recognition software uses image recognition
technology to recognize facial features and accurately identify an individual. Addi-
tionally, AR and VR applications use image recognition to accurately sense and display
overlay information [10]. Computer vision and image recognition are highly capable
technologies that enable a machine to understand the nuances and characteristics of
an image or video. These technologies have already been incorporated into many in-
dustries such as the automotive and medical industries, as well as facial recognition
applications, security, and AR and VR applications [11]. As technology advances, com-
puter vision and image recognition will continue to expand and develop, making
these technologies increasingly more powerful and useful.
The present focus of research and development in the field of machine automation
and decision-making is on robotics and autonomous systems (RAS). The aim is to cre-
ate machines that can perform tasks autonomously without the need for human
input. RAS have revolutionized the way many complex tasks and processes are com-
pleted, allowing organizations to automate labor-intensive tasks and to free up re-
sources for higher order activities. Research in this area has made great strides in
recent years, and it continues to remain an area of significant investment. One of the
key advancements made in the field of robotics was the introduction of autonomy.
This refers to a machine’s capacity to function without human input. Decisions are
made and tasks are completed based on the programmed logic, rather than by follow-
ing instructions from a person. Autonomous robots can sense their environment, ana-
lyze it, and react to it in order to plan and adjusts its actions accordingly [12]. This has
allowed for a wide range of robotics purposes, including surveillance, space explora-
tion, service robots, manufacturing, and more. Another major breakthrough has been
the development of AI, which allows robots to learn from their experiences and apply
what they have learned in order to complete tasks. AI allows robots to develop better
algorithms and methods to approach problems [13]. AI can also be combined with sen-
sors, image recognition, and other capabilities to allow robots to complete more com-
plex tasks. Robots are also becoming increasingly better at cooperating with people.
Human-robot collaboration, in which robots and people can work together to achieve
a goal, is becoming increasingly popular and is being used in many aspects of life [14].
This includes things like industrial robots used in the manufacturing sector, domestic
robots used in the home, and medical robots used in the healthcare environment.
RASs represent a remarkable technological achievement, and their possibilities are
endless. They will continue to have a profound impact on the way tasks are per-
formed and decisions made in the near future, and the opportunities are exciting.
8 Mahesh Maurya et al.
Expert systems, which are knowledge-based systems intended to mimic the decision-
making of human experts and deliver solutions in the form of recommendations,
models, and predictions, are one of the cutting-edge AI technologies. Expert systems
store and manage knowledge in a very specific and predetermined representation,
known as knowledge representation. Knowledge representation on expert systems is
usually formalized and displayed through ontologies, taxonomies, and frames, which
are structured according to the contextual data of the expert system. Ontologies are a
type of knowledge representation on expert systems composed of concepts, relation-
ships between them, and their properties that define and differentiate them from
each other. The concept of ontologies differentiates itself from taxonomies in that it
not only identifies hierarchical relationships between objects, but can also store the
characteristics that differentiate each object from the others and represent that
knowledge in the form of statements. An example of an ontology in expert systems is
an ontology that represents the characterization and classification of the objects expe-
rienced by the system. For example, an expert system for stock market trading might
have an ontology representing different types of stocks and their variation in parame-
ters such as price, volume, and volatility. A type of knowledge representation very
similar to ontologies is frames. Frames are graphical representations of objects that
use slots, which imply fixed characteristics, and the amount of slots vary depending
on the context. Frames are used in knowledge-based Systems because they provide a
comprehensive representation of individual concepts or objects, covering different
characteristics of them. An example of a frame within an expert system would be a
frame representing a stock or bond, which contains information about the stocks
price, volatility, yield, and market capitalization. Taxonomies are another type of
knowledge representation used in expert systems. Taxonomies are a way of repre-
senting knowledge by categorizing objects into groups. The taxonomy of an expert
system contain nodes which functions as a hierarchical structure that organizes con-
cepts into categories based on their similarity or shared characteristics. An example
of a taxonomy on an expert system could be a taxonomy that categorizes stocks by
their industry or sector into groups such as energy, materials, and technology [15].
The expert systems use knowledge representations such as ontologies, taxonomies,
and frames to efficiently organize and store information about objects experienced
by the system. Each of these representations has distinct characteristics and is used to
display different aspects of the knowledge stored by the system.
Machines and software agents can interact with their environment using reinforce-
ment learning (RL), a type of AI to learn how to optimize their reward. It is a branch
1 Introduction to artificial intelligence 9
The two most popular techniques to ML and AI are neural networks and ANNs. Neu-
ral networks are predicated on the notion that the brain is made up of a significant
number of interconnected neurons that cooperate to produce a desired result. ANNs
are created to replicate the biological structure and behavior of the neurons found in
the brain in computer code. A “back-propagation algorithm” is used in a neural net-
work to find patterns in the data and change weights and other parameters to achieve
the desired outcome. This outcome is often a prediction, such as the optimal course of
action in a certain circumstance. An input layer, hidden layers, and an output layer
make up a neural network. The data is processed using the nodes and connections
between them, which produce outputs according on the parameters. A wide range of
tasks can be carried out by ANNs. For example, ANNs are often used for face recogni-
tion, image classification, fraud detection and NLP. By making use of the vast amounts
of data available, ANNs can learn a variety of tasks and can continue to learn and
improve themselves over time. ANNs have a number of advantages over other ML al-
gorithms. For example, ANNs have greater flexibility and can be adapted to the type
of problems more easily. Additionally, ANNs can be trained more quickly and accu-
rately than other algorithms. However, ANNs do have some drawbacks. Training and
evaluating an ANN can require a great deal of information and layering in order to
achieve accurate results. Additionally, ANNs can suffer from overfitting, which is
when the algorithm fails to generalize and performs worse given new data. For ML
and AI, neural networks and ANNs are effective and practical methods. They offer a
10 Mahesh Maurya et al.
fantastic approach to swiftly and accurately analyze data and generate results [17].
While they can suffer from some issues, such as overfitting, there are ways to avoid
these problems. With the right training and evaluation, ANNs can provide an invalu-
able tool for many different tasks.
1.6 Challenges in AI
The biggest challenge in AI is to develop artificially intelligent systems that can effec-
tively combine the use of existing AI technologies with human knowledge and intui-
tion to make decisions and solve problems. Additionally, AI must be able to operate in
complex and unpredictable environments. Achieving these aims becomes more cru-
cial as AI continues to permeate our daily lives, both for the benefit of the AI develop-
ment community and for the society as a whole. One of the biggest AI challenges is
creating environments that are easy for both machines and humans to understand
and interact with. As AI progresses, developers need to not only develop smarter ma-
chines that are able to make decisions more effectively, but also create environments
in which machines can easily parse and comprehend the abundance of data available
today. In addition, AI must become better at understanding both spoken and written
language, enabling it to interact with humans in more natural conversation. AI must
also become better at reasoning and problem solving, as well as understanding and
learning from humans. All these capabilities should ideally be combined with a self-
regulating mechanism, enabling AI to improve itself.
even more important to consider. It is essential that businesses and organizations using
AI have a set of guidelines and policies in place to ensure ethical data practices, create
transparency and accountability throughout the AI process, and adhere to international
regulatory frameworks. The effects of implemented AI-based systems on vulnerable in-
dividuals or populations must also be taken into account [18]. To ensure ethical use of
AI, businesses should create and adhere to a set of values that decrease the potential
for bias or other unethical outcomes. AI holds the potential for far-reaching innovation
and progress, and its responsible development and deployment is of paramount con-
cern. For AI to work in the long term, such considerations must be taken seriously and
addressed directly.
Various areas, including healthcare, banking, and education use AI algorithms. They
have the ability to completely alter how we live and work. They do, however, carry a
chance of bias and unfairness. AI algorithms can replicate and even exacerbate preex-
isting biases in the data they are trained on since they are only as good as the input
they are given. This may result in arbitrary decisions regarding loan applications, job
applicants, and college admissions. In order to prevent bias from entering into AI al-
gorithms, data needs to be examined and preprocessed to ensure it is structured in
ways that do not inherently include bias. Data sets must be diversified to ensure the
algorithm is exposed to a variety of experiences and perspectives. There must also be
diversity in the people designing the algorithm and ethical considerations should be
built into the system. The algorithmic transparency must be maintained. The methods
of the algorithm should be easy to understand and available for scrutiny. Measures
must be put in place to ensure responsible adjustment for any bias that is found in
the decisions made. Finally, an open and continuous dialogue should be held between
AI systems and stakeholders to ensure a sense of fairness and trust in the system. The
potential of AI is immense and must be regulated to ensure that bias and fairness are
accounted for. With proper training and oversight, AI algorithms can provide better
and fairer decisions without sacrificing accuracy.
privacy and security. Governments must develop regulations and laws that protect indi-
viduals’ data from unwanted access and only allow verified and trusted organizations
to access such data. Organizations are responsible for following regulations and best
practices put in place by governments to ensure the highest level of data security and
privacy. They must also make sure that data collected is used for the specific purpose it
was collected for and that it is not exposed to unauthorized users. Last but not least,
people need to be aware of how their data is being used and make sure that critical
information like social security numbers and credit card information is not shared with
anyone. The algorithms used to handle and interpret data are one of the main privacy
and data security problems in AI. These algorithms can be used to find trends and
make judgments about people’s behavior and traits, putting the privacy of people’s per-
sonal information at stake. Algorithms should be examined closely and comply with
laws and regulations to ensure that no individual’s data is used without his/her permis-
sion [19]. While AI offers many potential benefits to society, governments, organiza-
tions, and individuals must take the necessary steps to ensure that data privacy and
security are not compromised. Governments must put in place regulations and organi-
zations must take proactive measures to make sure that all information collected is
used responsibly and securely. Finally, individuals must be aware of their data usage
and be vigilant against malicious actors who may be trying to take advantage of them.
As AI becomes more complicated, it is becoming more and more vital to worry about
the interpretability and explainability of AI systems. Understanding how an AI system
makes decisions is crucial for building trust in the system and its outcomes as compli-
cated AI systems become more prevalent. The ability of humans to comprehend the
inner workings and logic of AI systems is referred to as interpretability [20]. This is
important for providing an understanding of the problems the AI system is trying to
solve, and the reasoning and decisions it has made. Explainability, on the other hand,
relates to the ability of AI systems to provide an explanation of their behavior. Ex-
plainability includes providing meaningful explanations of why a system behaves the
way it does and should ideally provide insight into the reasoning and logic of the sys-
tem. An essential topic of AI research is the creation of interpretable and explicable
AI systems [18]. AI researchers will need to make use of computer-readable represen-
tations and comprehensible methods like symbolic and statistical AI, NLP, and ML in
order to develop these systems. While AI systems are currently incredibly powerful,
their lack of interpretability and explainability is preventing them from being used in
many industries. By creating interpretable and explainable AI systems, humans will
be able to understand and trust how these systems make decisions, paving the way
for broader applications of AI.
1 Introduction to artificial intelligence 13
AI has been developing quickly in recent years, enabling machines to take on an in-
creasing number of tasks that were previously performed by people. Such develop-
ments have prompted a number of inquiries, the most important of which is whether
and to what degree AI will result in job displacement. According to some analysts, AI
may result in the loss of millions of jobs because many functions can be done by ma-
chines more effectively and cheaply than by humans. As AI becomes more advanced,
redundant or low-skill occupations may be phased out, as machines become more via-
ble in such roles. For example, customer service and administrative support roles are
increasingly being automated with AI-driven machines and chatbots that provide
basic customer service interactions. Other experts argue that while some jobs may
disappear, AI will actually lead to the increase in new types of jobs. As AI proliferates,
it will create an increasing demand for experts who understand how it works, and
how it can be used to advance a business. Additionally, AI systems will need to be
maintained, monitored, and kept updated, which requires a dedicated and knowl-
edgeable workforce [21]. Still, it is certain that AI will leave an indelible impact on
jobs and how works gets done in the future. It is therefore incumbent on governments
and businesses to plan and prepare for the job displacement that is expected to occur.
Governments must implement policies that encourage training and re-education for
displaced workers, increasing their employability in the AI economy. Businesses need
to take responsibility for providing current employees with the skills needed to suc-
ceed in the AI workplace. There is little doubt that AI has the ability to bring about
employment displacement as well as significant growth and prosperity. By being agile
and preparing for the changing job market, we can ensure we maximize the benefits
and mitigate the risks of further advancing AI.
The development of AI systems is creating various policy and regulatory challenges for
governments and organizations. As AI algorithms become increasingly complex and au-
tonomous, policymakers must grapple with regulatory questions related to the design,
use, and effects of the technology. This essay will examine the key regulatory and policy
challenges posed by AI systems through an analysis of issues relating to privacy, liabil-
ity, trust, security, and ethical considerations. Privacy is perhaps the most important
regulatory concern posed by the increasing use of AI algorithms. As AI algorithms be-
come more widely used, they rely on access to large amounts of data. People frequently
submit this data, maybe unaware of the consequences of doing so when supplying it to
AI systems. To ensure that individuals maintain control over their personal data, ade-
quate privacy protections should be established [22]. This may involve constraint on the
type of data that can be collected, limits on how long data can be stored, and limits on
14 Mahesh Maurya et al.
how personal data can be used. Liability is also an important regulatory issue in the
context of AI algorithms. As algorithms become increasingly autonomous, it is possible
that they will be held responsible for their own decisions or actions. This raises ques-
tions over who should be held liable in case of an algorithm failure or mistake. This
may involve issues such as whether algorithm developers or users should be liable or
whether liability should be assumed by the government. Trust is another important pol-
icy consideration when it comes to AI systems. Trust in AI algorithms is essential for
their use, as the public must be able to rely on the accuracy and reliability of the deci-
sions made by the algorithms. To ensure trust, it will be necessary for AI systems to be
governed by a framework of accountability and transparency. This may involve devel-
oping technical standards for AI systems, as well as certification and audits for develop-
ers and users of the technology. Security is also an important regulatory concern, as AI
algorithms have the potential to be used maliciously or for malicious ends. The security
of AI systems must be assured through rigorous security measures. These may involve
controls over the implementation and use of the algorithms, as well as measures to pre-
vent malicious manipulation of the data used for training the algorithms. The ethical
considerations are another key regulatory challenge posed by AI algorithms. AI algo-
rithms must be designed in such a way that they respect fundamental ethical principles
and values. This may involve, for example, ensuring that the algorithms are not used to
discriminate against certain groups or that they adhere to appropriate privacy protec-
tions [23–25]. A variety of regulatory and policy concerns are brought about by AI sys-
tems. These include concerns with respect to privacy, liability, trust, security, and
morality. To guarantee that AI algorithms are used safely, responsibly, and in a way
that respects people’s rights and values, policymakers must address these issues.
The popularity and advancement of AI have risen dramatically in recent years. AI de-
scribes a machine’s capacity to learn and reason similarly to a person. This technology
has the potential to completely transform people’s lives all around the world and can
be applied in a range of contexts, including industrial automation and medical diagno-
sis. Although AI may have certain advantages, it also contains some risks and has the
potential to have a big impact on society. One of AI’s most significant effects is that it
allows machines to perform tasks that were previously thought to be performed only
by humans. As AI technology advances, machines can increasingly be used to replace
human labor. This could lead to massive job losses and an increase in income inequal-
ity, for example. Moreover, AI machines may be vulnerable to attack from malicious
actors, putting important data and systems at risk. Another major concern is the ethical
implications of AI technology [26]. AI is being used to make decisions that affect human
lives, such as medical diagnoses and loan approvals. The decisions made by AI systems
may be based on biased or inaccurate data, creating ethical and social dilemmas. The
1 Introduction to artificial intelligence 15
potential for abuse and misuse of AI technology is also a major concern [27]. The effects
of AI technology over the long run must be taken into account. AI machines have the
ability to surpass human intelligence, which could result in technological advancements
that are difficult to be foreseen. This could have a profound effect on human society,
potentially leading to the displacement of millions of people, or the creation of new clas-
ses of people who possess great power and wealth [28]. The AI technology has a variety
of potential impacts on society. It can be used to replace human labor, create ethical
dilemmas, and potentially lead to technological advances that could drastically change
the way people interact with each other. As AI technology continues to advance and
become more prevalent, it is important to consider these potential societal impacts so
that they can be better managed and addressed.
1.7 Conclusion
AI has the potential to revolutionize the world with its numerous innovative applica-
tions. This technology has grown at a rapid rate and can be seen as a major stepping
stone in the future of technology. AI is capable of problem-solving, decision-making,
and finding subtle patterns in information that humans cannot detect. It is expected
to have a wide range of applications in the future, ranging from healthcare, finance,
transportation, security, and many more. Using NLP technology to help doctors better
comprehend the problems their patients are facing is a significant application of AI in
the healthcare industry. AI can help doctors make better judgments and deliver more
individualized care by analyzing the talks between the patient and the doctor to gain
insights into the patient’s mental health, exhaustion, or other difficulties. AI is also
being used in the finance sector to evaluate data more effectively. Applications like
fraud detection, risk analysis, portfolio management, and personal finance manage-
ment fall under this category. Financial institutions may make judgments more rap-
idly by using AI to spot trends in data. AI is being used in the transportation industry
as well to enable autonomous vehicles to drive safely and efficiently in any situation.
This technology can detect and react to changes in the environment quickly and re-
spond appropriately. AI can also be used to optimize traffic flow and public transpor-
tation around cities to reduce congestion. AI is also becoming essential to enhancing
security. AI can be used in criminal identification, facial recognition, and military ap-
plications. AI can detect suspicious activity quickly and provide more accurate detec-
tion than humans are able to. AI is important because it enables us to process large
amounts of data and gain insights that were not possible before. It can also provide
solutions to problems faster and more accurately than humans can. We must keep
innovating in this area if we are to gain from AI, which will play a significant role in
our future. The question of how AI will affect humanity’s future is still up for debate.
AI has already made significant breakthroughs in the world of technology and robot-
16 Mahesh Maurya et al.
ics, allowing us to automate and optimize processes and operations, as well as make
predictions and automate decisions. Ultimately, the future of AI will depend on how it
is used and applied. If AI is used responsibly, with a focus on creating positive and
beneficial solutions for society, it can help build a better future for humanity. To
bring responsible AI into action, governments, businesses, and AI developers need to
collaborate, with an emphasis on transparency, accountability, and ethical principles.
By regularly researching the use of AI in the real world and adjusting policies and
regulations accordingly, we can build an AI-driven future that puts humans first.
The emerging trends and advancements in AI are revolutionizing the way we ap-
proach data and computation. In recent years, AI has progressed beyond traditional
AI techniques, such as ML, to include deep learning algorithms, NLP, and robotics. AI
is used for a variety of tasks from image recognition to voice recognition to robotic
process automation. Deep learning algorithms debuted with great success, as it gave
AI the ability to produce accurate results without needing vast amounts of training
data. NLP saw success with its ability to parse out meaning from large quantities of
text and audio data. Automation saw progress with the ability to automate mundane
and repetitive tasks without human intervention. Today, AI is rapidly advancing to
meet the growing demand for more efficient and effective data processing. AI-based
software can identify patterns in large data sets, predict customer behavior, and even
autonomously control robots. The rapidly developing technology of AI is also being
used to create AR and VR experiences. AI is being applied to a wide variety of areas,
ranging from healthcare to banking to retail. As AI continues to advance, the possibili-
ties for its implementation will expand rapidly. The advancements in AI have made it
1 Introduction to artificial intelligence 17
easier than ever for businesses and organizations to access powerful data analysis
tools. AI-driven insights can drive performance and productivity improvements, al-
lowing organizations to discover previously hidden trends and insights. Cloud-based
AI solutions are also becoming increasingly popular, giving organizations the ability
to quickly implement AI solutions without significant upfront investment. The ad-
vancements in AI-related technologies are continually driving progress. AI-based ro-
bots are becoming smarter and more capable every day, and advancements in ML
and NLP are making it easier for AI systems to understand their surroundings and
interact with humans. In addition, AI-driven systems are increasingly able to replace
humans in a variety of roles, driving efficiency and cost savings. As AI continues to
evolve, its applications will become even more diverse. AI-driven solutions that auto-
mate mundane tasks have the potential to reduce human workload, while AI-driven
insights can inform decision-making and improve results. Regardless of the industry
or field, AI is sure to continue to drive innovation in the years to come.
The emergence of AI in recent years has been the most significant technology develop-
ment since the invention of the internet. AI is already having an incredible impact on
multiple industries and is projected to revolutionize the way our society functions. AI’s
potential for improvement and disruption is unprecedented as its applications are end-
less. The most direct and visible impact of AI is on industries, such as healthcare, fi-
18 Mahesh Maurya et al.
nance, and transportation. For example, AI has enabled medical breakthroughs, such as
more accurate diagnoses and treatments, which have enabled the healthcare industry
to move faster and more efficiently than ever before. AI has also had an impact on fi-
nance, with the development of algorithmic trading, which has enabled banks and
asset managers to process trades at a much faster rate. Additionally, AI has been used
to optimize transportation services, such as ride-sharing apps, to better manage fleets
of vehicles and optimize route planning for drivers. AI is also transforming the way we
interact with technology. Through AI-driven NLP technology, users can increasingly in-
teract with machines in a more natural and intuitive manner. This technology has
enabled more effective task automation and improved customer service, which is ex-
pected to play a major role in economic sectors such as retail, food services, and hospi-
tality. AI is also likely to have a profound impact on society, potentially replacing
human labor and ushering in a new era of automation. It is expected that much of the
busy work involved in everyday life, such as the automation of manufacturing, will be
replaced by AI. This automation is expected to reduce errors and improve the overall
efficiency of organizations. In addition, AI-driven analytics are expected to revolution-
ize the way we interact with businesses, enabling consumers to access more accurate
and personalized services. Ultimately, the impact of AI on industries and society is far-
reaching and potentially revolutionary. AI has already enabled major improvements in
industries such as healthcare, finance, and transportation, and is expected to reshape
the way we interact with technology and with each other. We are only just beginning to
understand the potential of AI and the opportunities that it presents, but it is clear that
it will have a significant impact on our lives in the years to come.
free of algorithmic bias or other discriminatory effects. Ethical considerations and re-
sponsible AI practices should be an integral part of AI development. Ensuring that
ethical considerations are taken into account when building AI systems and that they
adhere to responsible AI practices will help ensure that the technology is used safely
and without detriment to individuals.
References
[1] Ongsulee, P. (2017, November). Artificial intelligence, machine learning and deep learning. In 2017
15th International Conference on ICT and Knowledge Engineering (ICT&KE) (pp. 1–6). IEEE.
[2] Norvig, P. R., & Intelligence, S. A. (2002). A modern approach. Upper Saddle River, NJ, USA: Prentice
Hall; Rani, M., Nayak, R., & Vyas, O. P. (2015). An ontology-based adaptive personalized e-learning
system, assisted by software agents on cloud storage. In Knowledge-Based Systems Vol. 90
(pp. 33–48).
[3] Alimadadi, A., Aryal, S., Manandhar, I., Munroe, P. B., Joe, B., & Cheng, X. (2020). Artificial
intelligence and machine learning to fight COVID-19. Physiological Genomics, 52(4), 200–202.
[4] Awasthi, S., Naruka, M. S., Yadav, S. P., & De Albuquerque, V. H. C. (Eds.). (2023). AI and IoT-based
intelligent health care, & sanitation. India: Bentham Science Publishers.
[5] Brewka, G. (1996). Artificial intelligence – a modern approach by Stuart Russell and Peter Norvig,
Prentice Hall. Series in Artificial Intelligence, Englewood Cliffs, NJ. The Knowledge Engineering
Review, 11(1), 78–79.
[6] Gerke, S., Babic, B., Evgeniou, T., & Cohen, I. G. (2020). The need for a system view to regulate
artificial intelligence/machine learning-based software as medical device. NPJ Digital Medicine,
3(1), 53.
[7] Chauhan, S. S., Yadav, S. P., Awashthi, S., & Naruka, M. S. (2023). Sentimental analysis using cloud
dictionary and machine learning approach. In Cloud-based Intelligent Informative Engineering for
Society 5.0 (pp. 157–169). India: Chapman and Hall/CRC.
[8] Li, Y., Kumar, R., Lasecki, W. S., & Hilliges, O. (2020, April). Artificial intelligence for HCI: A modern
approach. In Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing
Systems (pp. 1–8).
[9] Hyder, Z., Siau, K., & Nah, F. (2019). Artificial intelligence, machine learning, and autonomous
technologies in mining industry. Journal of Database Management (JDM), 30(2), 67–79.
[10] Siau, K., & Wang, W. (2018). Building trust in artificial intelligence, machine learning, and robotics.
Cutter Business Technology Journal, 31(2), 47–53.
[11] Krishna, A. B., Tanveer, A., Bhagirath, P. V., & Gannepalli, A. (2020). Role of artificial intelligence in
diagnostic oral pathology-A modern approach. Journal of Oral and Maxillofacial Pathology: JOMFP,
24(1), 152.
[12] Das, S., Dey, A., Pal, A., & Roy, N. (2015). Applications of artificial intelligence in machine learning:
Review and prospect. International Journal of Computer Applications, 115(9), 31–41.
[13] Saravanan, V., Rajeshwari, S., & Jayashree, P. (2013). SECURITY issues in protecting computers and
maintenance. Journal of Global Research in Computer Science, 4(1), 55–58.
[14] Pollack, M. E. (1995). Artificial intelligence – A modern approach – A review. Ai Magazine, 16(3),
73–73.
[15] Khanzode, K. C. A., & Sarode, R. D. (2020). Advantages and disadvantages of artificial intelligence
and machine learning: A literature review. International Journal of Library and Information Science
(IJLIS), 9(1), 3.
20 Mahesh Maurya et al.
[16] Woschank, M., Rauch, E., & Zsifkovits, H. (2020). A review of further directions for artificial
intelligence, machine learning, and deep learning in smart logistics. Sustainability, 12(9), 3760.
[17] Yadav, S. P., & Yadav, S. (2019). Mathematical implementation of fusion of medical images in
continuous wavelet domain. Journal of Advanced Research in Dynamical and Control System, 10(10),
45–54.
[18] Sultan, A. S., Elgharib, M. A., Tavares, T., Jessri, M., & Basile, J. R. (2020). The use of artificial
intelligence, machine learning and deep learning in oncologic histopathology. Journal of Oral
Pathology and Medicine, 49(9), 849–856.
[19] Bini, S. A. (2018). Artificial intelligence, machine learning, deep learning, and cognitive computing:
what do these terms mean and how will they impact health care?. The Journal of Arthroplasty, 33(8),
2358–2361.
[20] Žarković, M., & Stojković, Z. (2017). Analysis of artificial intelligence expert systems for power
transformer condition monitoring and diagnostics. Electric Power Systems Research, 149 125–136.
[21] Asemi, A., Ko, A., & Nowkarizi, M. (2020). Intelligent libraries: a review on expert systems, artificial
intelligence, and robot. Library Hi Tech, 39(2), 412–434.
[22] Jakhar, D., & Kaur, I. (2020). Artificial intelligence, machine learning and deep learning: Definitions
and differences. Clinical and Experimental Dermatology, 45(1), 131–132.
[23] Jones, L. D., Golan, D., Hanna, S. A., & Ramachandran, M. (2018). Artificial intelligence, machine
learning and the evolution of healthcare: A bright future or cause for concern? Bone and Joint
Research, 7(3), 223–225.
[24] Nithya, C., & Saravanan, V. (2018). A study of machine learning techniques in data mining.
International Scientific Refereed Research Journal, 1 31–38.
[25] Aruna, R. D., Surendran, S. D., Yuvaraj, N. D., & Debtera, B. (2022). An enhancement on
convolutional artificial intelligent based diagnosis for skin disease using nanotechnology sensors.
Computational Intelligence and Neuroscience, 2022, 1–13.
[26] Silvia Priscila, S., Sathish Kumar, C., Manikandan, R., Yuvaraj, N., & Ramkumar, M. (2022, February).
Interactive artificial neural network model for UX design. In Arulmurugan Ramu, Chow Chee Onn &
M.G. Sumithra (Eds.). International Conference on Computing, Communication, Electrical and
Biomedical Systems (pp. 277–284). Cham: Springer International Publishing.
[27] Adams, J. A. (2001). Multiagent systems: A modern approach to distributed artificial intelligence. AI
Magazine, 22(2), 105–105.
[28] Ghahramani, Z. (2015). Probabilistic machine learning and artificial intelligence. Nature, 521(7553),
452–459.
Shaik Mohammad Rafee, M. Prasad, M. Sunil Kumar,
Balamurugan Easwaran
2 AI technologies, tools, and industrial
use cases
Abstract: Artificial intelligence (AI) technology refers to various technologies and tools
used in the development of systems that can carry out tasks that normally require
human intelligence. AI systems can process large datasets and communicate in natural
languages, taking decisions, recognizing objects, detecting anomalies, and managing
data to provide personalized recommendations. AI technologies can be used in almost
any field, including healthcare, finance, e-commerce, marketing, robotics, Internet of
things, and more. Popular AI technologies and tools include computer vision, natural
language processing (NLP), machine learning (ML), deep learning (DL), sentiment analy-
sis, voice recognition, and robotics. Computer vision is used to identify objects in pic-
tures or videos. NLP is used to understand language. ML is used for data modeling and
to find patterns in data. DL techniques are used to process large amounts of data and
automate decisions. Sentiment analysis is used to analyze customer feedback. Voice rec-
ognition is used to recognize and understand spoken words. Robotics technology is
used to build and program robots. Businesses are increasingly using AI technologies
and tools to improve customer experience, increase efficiency, automate tasks, increase
scalability, and provide personalized recommendations and services.
https://doi.org/10.1515/9783111323749-002
22 Shaik Mohammad Rafee et al.
Supervised learning algorithms are algorithms that use labeled data to learn from.
Labeled data is data that has been marked or identified in some way, making it eas-
ier for the algorithm to understand the patterns that exist in the data. Supervised
learning algorithms can be used to solve a wide range of tasks, such as predicting
stock market trends, determining the customer segments that are likely to purchase
new products, or classifying images and other types of data. By using labeled data,
supervised learning algorithms are able to build models that are better able to make
24 Shaik Mohammad Rafee et al.
Unsupervised learning algorithms are useful for exploring data, uncovering hidden
patterns, and making predictions. They can be used in many applications, such as cus-
tomer segmentation, disease diagnosis, fraud detection, and more. As ML algorithms
become increasingly sophisticated, the potential uses for unsupervised learning algo-
rithms are expected to grow [11].
RL is a type of ML algorithm that enables artificial agents to learn from their environ-
ment and optimize their actions by trial and error. Unlike traditional supervised
learning algorithms, RL algorithms learn from rewards and punishments, allowing
the agent to learn more quickly and adapt to changing conditions. RL algorithms are
composed of several components, including an agent, a learning environment, a re-
ward function, and a set of actions. The agent learns by observing the environment
and then deciding which action to take based on the rewards and punishments it re-
ceives. The environment provides feedback to the agent in the form of rewards or
punishments, which influence the learning process. When the agent makes a “correct”
decision, it is rewarded with a positive reward. If it makes a “wrong” decision, it is
punished with a negative reward [12]. This reward system encourages the agent to
learn a “correct” policy that maximizes the rewards it receives. The goal of RL algo-
rithms is to maximize the long-term rewards an agent receives. As the agent takes
actions, it accumulates rewards in a cumulative fashion so the longer it engages in the
task, the higher the rewards it receives. Many different RL algorithms exist, each with
a slightly different approach or a set of parameters. Examples of popular RL algo-
rithms include Q-learning, SARSA, and evolutionary algorithms. No matter which al-
gorithm is used, RL can provide powerful solutions to complex problems such as
robotic navigation, finance, and games. It can also be used in simple problems such as
in navigation, where the agent needs to find the most efficient path to an endpoint
[13]. Regardless of the type of problem, RL algorithms are a powerful tool that can
help agents to learn a “correct” policy, while maximizing rewards.
Deep learning (DL) algorithms are a branch of AI algorithms modeled after the struc-
ture of the human brain. DL algorithms are composed of artificial neural networks
and are capable of unsupervised learning, supervised learning, and RL. They are used
mainly for tasks such as image recognition, NLP, and self-driving cars. The typical
structure of a DL algorithm consists of several layers of processing units, which are
usually called neurons. Each neuron within the network passes data to the nest neu-
ron, and all of these neurons interact with each other and with the original input
26 Shaik Mohammad Rafee et al.
data. In each layer, the neurons connect to each other to form an artificial neural net-
work. Optimizing the weights of the connections between the neurons allows the
model to generalize useful patterns in data [14]. In some cases, DL algorithms are
trained using millions of examples of data. For example, a DL algorithm, trained on a
dataset of millions of images of cats and dogs, can be used for a task like image recog-
nition. The algorithms build up more and more complex levels of information from
the dataset until it can recognize and distinguish cats and dogs. DL algorithms are be-
coming increasingly powerful as the number of neurons and layers within the net-
works increase, and are now being used in many real-world applications, such as
medical diagnosis, drug discovery, and automated language translation [15]. With the
increasing availability of extremely large datasets, DL algorithms will become even
more powerful and will eventually become a major part of our lives.
generalization. By leveraging the diversity of its weak models, it can create a more
robust and accurate model. It is becoming increasingly popular and is now commonly
used in a variety of applications such as object recognition and Neuroevolution.
2.4.1 TensorFlow
2.4.2 PyTorch
PyTorch is an open-source ML library for Python, used primarily for DL. It is devel-
oped by Facebook’s AI Research Group and is designed to be fast and flexible. PyTorch
allows users to easily and quickly create and deploy DL models on a variety of devi-
ces, including GPUs, CPUs, and even mobiles. PyTorch supports automatic differentia-
tion, a technique that allows the library to compute gradients and backpropagation.
This means that users can easily and quickly train DL models, making it ideal for ap-
plications such as computer vision and NLP. PyTorch also offers native support for
popular DL frameworks such as TensorFlow, Caffe2, and Theano. PyTorch also offers
several advantages over other DL libraries. For example, it offers high performance
with both CPUs and GPUs [22]. This allows users to quickly develop and deploy DL
models to run on a variety of devices. Additionally, the library is designed to be user-
friendly and easy to learn, making it suitable for both experienced and novice users.
PyTorch offers excellent support for distributed training, thanks to its distributed
package. This allows users to train models across multiple machines and GPUs, which
can potentially help reduce training time. This is especially beneficial for training
large models, which might otherwise take a long time to train on a single machine.
2 AI technologies, tools, and industrial use cases 29
2.4.3 Keras
2.4.4 Scikit-learn
the optimal parameters are identified, and that overfitting is avoided when building
models. Moreover, its practical tutorials allow developers to quickly become familiar
with the library and its various features. The Scikit-learn is an incredibly valuable
tool for data scientists and ML professionals. It provides a wide range of robust algo-
rithms, powerful optimization techniques, and intuitive visuals for gaining insights.
And with its user-friendly interface, it makes ML tasks easier, faster, and more
reliable.
2.4.5 Theano
Theano is an open-source computing library for Python, developed at the LISA lab at
the University of Montreal. Theano is designed specifically to handle complicated nu-
merical calculations and to optimize code efficiently when multiple CPU cores and
GPUs are available. Theano can be used for a wide range of applications, such as ML,
computer vision, and NLP. At its core, Theano provides a basic set of functions and
operations that handle complex mathematical problems and reduce problems to their
most basic form. This enables ML models to run faster and more efficiently, as Theano
performs computations for algorithms in parallel [25]. Theano also simplifies the de-
velopment of DL networks by allowing network layers to be built quickly and easily.
DL networks require complex calculations to determine the weights and biases of
each layer. Theano can optimize these calculations and speed up learning. Theano
also supports distributed computing, which means that it can be used on a range of
different computers and machines to speed up computations even further. Other pow-
erful DL libraries, such as TensorFlow, have adopted Theano’s distributed computing
architecture for their own libraries. Theano offers debugging tools and data visualiza-
tions to aid in debugging, and help understand the model and its performance. This
allows developers to ensure they are making progress in their DL models and pro-
vides them with useful insights into how the model is being trained. Theano is a pow-
erful Python library for developing DL models, optimizing complex mathematical
equations and performing distributed computing. It is an invaluable tool for any ML
enthusiast.
2.4.6 Caffe
The Caffe tool is an open-source AI software library, designed to support DL and deep
RL models in ML applications. It is a framework for complex modeling and design of
various DL models. Developed by a team at UC Berkeley’s AMPLab and funded by the
Berkeley AI Research (BAIR) project, the Caffe tool is the brainchild of Professor Yann
LeCun and Professor Berkeley AI Research Group and is maintained by the Berkeley
Vision and Learning Center. The Caffe tool is written primarily in C++ and is made
2 AI technologies, tools, and industrial use cases 31
available under the BSD 2-Clause license. It consists of two libraries: Caffe and Caffe2.
The Caffe library provides a general interface for DL networks. Through the use of
this library, users can easily create, modify, and deploy neural networks without hav-
ing to worry about manually coding up the model architecture and training algo-
rithms [26]. Caffe2 is an extension to Caffe that includes a number of pre-trained
models as well as supports distributed training. This makes it easier for users to ex-
periment with various DL models and architectures. The library also provides a set of
helpful APIs and tools to facilitate DL research and applications. Using Caffe, users
can quickly identify objects and features from images and videos, and can also gener-
ate predictions about future data set. By making use of this library, researchers and
developers can easily build and deploy complex neural networks and DL models.
2.4.7 MXNet
stand human language, NLP uses advanced tools such as tokenizers, part-of-speech
taggers, and semantic role labelers. These tools are used to identify and then analyze
the meanings of words and phrases in a given sentence. It also enables computers to
understand the context in which a sentence is used. NLP also relies on algorithms
such as support vector machines, decision trees, and neural networks for natural lan-
guage understanding. These algorithms are trained using large amounts of data so
that they can identify patterns and interpret natural language [28]. In addition, NLP
also incorporates methods such as ontologies to organize complex data, and fuzzy
logic to interpret various levels of ambiguity. NLP contains several powerful tools that
make it possible to interact with computers in natural language. These tools include
natural language parsing, text summarization, and text-to-speech conversion. Natural
language parsing is used to identify and understand the meaning of a given sentence.
Text summarization enables computers to understand and summarize long texts to a
shorter form. Text-to-speech conversion is used to transform text into voice messages.
NLP tools are becoming increasingly popular and are used in many areas such as AI,
e-commerce, marketing, government, and healthcare. Financial institutions, in partic-
ular, have largely benefited from the use of this technology in the form of automated
customer service, fraud prevention, and credit risk assessment. The development of
NLP tools has enabled automated customer service to answer customer queries
quickly and accurately. Fraud prevention services detect unusual behavior and pre-
vent fraudulent transactions. NLP tools represent a revolutionary advancement in the
field of AI. Automation has enabled companies to better understand the language of
their customers, detect fraud, and manage risk. NLP tools have been used to develop
better customer service, research, and analysis. It is no wonder that this technology is
revolutionizing the way many businesses interact with their customers.
Tokenization and text preprocessing are essential components of NLP tools. Tokeniza-
tion is the process of dividing a large body of text into smaller tokens such as words,
subwords, characters, phrases, and sentences. It is the fundamental step for any NLP
task as it makes the text ready to be processed by further algorithms. Text preprocess-
ing consists of several operations to clean the collected text data and make it suitable
for the application. These operations include removing special characters, whitespa-
ces, HTML tags, stop words, non-English words, punctuation marks, etc. It depends on
the application and the use case of the NLP model for which the text preprocessing
step may be added, reduced, or skipped. Tokenization is a critical step as incorrect
tokenization can lead to incorrect results; thus, special attention should be paid to
this step. After tokenization, preprocessing operations help in reducing the noise pres-
ent in the text and make it more suitable for further processing [29]. They can be used
to normalize the text, convert words and abbreviations, lemmatize and stem the
2 AI technologies, tools, and industrial use cases 33
words, etc. Corrupt or incomplete text data causes problems in training and evalua-
tion of the NLP model; therefore, it is important to clean them before training, to get
accurate results. Both tokenization and text preprocessing are important NLP tools
with many of its applications such as text classification, sentiment analysis, text sum-
marization, machine translation, etc. They help in optimizing the data by removing
unnecessary information and make it more suitable for NLP-related tasks. Therefore,
they are often used as essential preprocessing steps in NLP models.
Named entity recognition (NER), also known as entity extraction, is an important ele-
ment of NLP that identifies and categorizes specific items mentioned in text-based
documents, such as organizations, locations, time mentions, and quantity mentions.
This information can be used to better understand context, text relationships, and
search query intent. Examples of NER machines would include search engines catego-
rizing names, topics, products, and more in order to provide more informative search
results. NER has a long history and has been studied by many NLP researchers. In the
early days, researchers had to manually annotate data and apply rule-based systems
in order to identify certain entities. Modern approaches to NER are mostly focused on
DL-based solutions such as recurrent neural networks, long short-term memory net-
works, and convolutional neural networks as well as legacy-style systems such as sup-
port vector machines (SVM). Generally, these solutions can learn from annotated data
that had been previously collected or labeled, and can extract information from unla-
beled data [30]. NER is mainly used to extract names, dates, email addresses, locations,
etc. from text. It is becoming increasingly popular in a variety of applications such as
text summarization, opinion mining, and question answering. For instance, NER can
be used to extract the names of people and organizations from news articles in order
to summarize the article in a brief manner. Additionally, NER can be used to build
retargeting models and can aid in option mining in order to identify what people are
saying about products or services. NER is a valuable technology that is constantly
being improved and refined in order to provide better accuracy and more efficient
solutions. It has been used by various industries such as healthcare, finance, and secu-
rity, and is always making new advances and helping facilitate the processing of text
in a wide range of NLP applications.
Sentiment analysis is the process of determining the emotional tone behind a series of
words and is used to gain an understanding of the attitudes, opinions, and emotions
expressed within an online mention. It is commonly used to measure the sentiment of
34 Shaik Mohammad Rafee et al.
customer feedback, reviews, tweets, and other social media content in order to gain
insights into customer experiences. It can also be used to gain insights into what peo-
ple think about products, services, and more. This type of analysis enables businesses
to track the level of customer satisfaction and identify areas for improvement [31]. By
analyzing sentiment, organizations can better understand how customers perceive
their offerings and adjust their strategies in order to maximize customer satisfaction.
Text classification and clustering are two of the most popular methods of text mining.
Text classification is the process of assigning categories or labels to written documents,
based on their content. It is used to enhance the accuracy of web search engines, cate-
gorize documents, and to cluster similar documents together. Text clustering is the pro-
cess of assigning clusters of similar documents together. It is a mixture of unsupervised
and supervised ML algorithms. The main goal of text clustering is to identify and cate-
gorize data points, usually words and sentences, into meaningful groups. This is an im-
portant tool for analyzing large amounts of data with minimal effort. Clustering can
help reveal patterns and trends in documents as well as identify groups of related
words. Clustering also helps to reduce the noise and sparsity of the data, and allows
researchers to more easily identify the underlying structure of documents. Both meth-
ods are used for understanding and analyzing textual data. The use of text classification
and clustering determines the similarities and patterns of each text document. With
text classification, the goal is to assign labels and categories to documents and groups of
documents, while with text clustering, the goal is to group together documents that are
similar in content. Text classification and clustering are both powerful methods for un-
derstanding and analyzing textual data in both computational linguistics and NLP.
algorithms, and other statistical techniques. The data is then used in the application
of language modeling and generation tasks, such as language translation, question an-
swering, summarization, and summarization detection. There are several types of lan-
guage models that can be used for language modeling and generation. The most
common are RNN, CNN, LSTM, and GRU networks. RNNs are the most basic type of
recurrent neural networks that use temporal information to capture the structure of
the language. CNN models are used for detecting local patterns in text and for se-
quence-to-sequence learning tasks. LSTM models, on the other hand, add a memory
factor to perform tasks such as text classification and sentence completion. Finally,
GRU networks are the most advanced type of language modeling and generation net-
works, and they can perform tasks such as text summarization and text generation.
The data used in the language modeling and generation process can be varied. It can
include text documents, emails, webpages, audio recordings, and even images. This is
important because it ensures that the model can accurately capture the structure of
the data and generate responses.
Machine translation is a type of NLP tool that enables automated translation between
two languages. It is used mainly in commercial applications and often provides a way
of translating large amounts of content quickly and accurately. Machine translation is
useful in a variety of areas, such as translating documents, web content, and conver-
sation. Businesses and organizations abroad can translate customer and employee
emails or texts, helpdesk requests, product descriptions, and other web content
quickly and precisely; and research organizations can use machine translation for ac-
curate information gathering from large numbers of local documents. One of the key
advantages of machine translation is increased speed, as software can handle large
quantities of content quickly in comparison to manual translation. It can translate
hundreds of documents in a fraction of the time it would take human translators,
meaning businesses can serve their customers more efficiently and keep up with the
competition. However, machine translation is still limited by the language used and
context in which it is being used. Machines do not have the ability to understand the
nuances of language in the same way humans can, so it is important to ensure you
use a reputable provider of machine translation, and make sure the scope and context
of the documents is fully understood. Machine translation is a valuable tool for com-
panies and organizations that need to quickly and accurately translate large amounts
of content and documents in multiple languages. While machine translation does in-
volve some limitations, the benefits outweigh the costs and the resulting speed and
accuracy can improve customer service, market efficiency, and growth.
36 Shaik Mohammad Rafee et al.
Image preprocessing and augmentation are two key steps in the field of computer vi-
sion. Preprocessing typically refers to a set of techniques used to prepare an image
for further processing, such as feature extraction, object detection, or content classifi-
cation. Image augmentation, on the other hand, is a process of creating a new version
of an existing image by applying certain procedures such as shifting, flipping, or resiz-
ing. Preprocessing involves applying mathematical algorithms and other modifica-
tions to an image with the aim of improving its performance in further stages of the
imaging process. Preprocessing often includes correcting for any anomalies in the
image, adjusting color balance, sharpening edges, and increasing contrast or satura-
tion. This is done to make the image more accurate and consistent, and improve its
visual appeal. Augmentation, in contrast, is the process of creating new images from
existing ones. By applying techniques such as shifting, scaling, zooming, or flipping,
new images can be created from existing images or without any distortion or impact
2 AI technologies, tools, and industrial use cases 37
to the original image. This process is widely used in ML and DL to increase the quan-
tity of training data sets as well as to increase the diversity of each set. Image prepro-
cessing and augmentation are essential steps for any imaging project. The former
allows for minor adjustments to existing content, while the latter stimulates the crea-
tion of new image content. By properly executing these two stages, images can be im-
proved or expanded in a relatively easy and cost-effective manner.
Object detection and tracking are two important types of computer vision tasks that
have been widely adopted in a variety of applications, such as surveillance, automo-
tive, robotics, and entertainment. Object detection is the process of locating and recog-
nizing the presence of objects in an image or video. This could be as simple as
detecting faces or identifying objects present in the scene. On the other hand, tracking
involves the process of locating and tracking objects in a video or image sequence,
such as following a person in a video or tracking the movement of a particular object.
Object detection and tracking have become increasingly popular over the years as
they improve the accuracy and efficiency of computer vision applications. Object de-
tection algorithms allow for improved accuracy in recognizing the presence and type
of objects in video and images, which can then be used for tracking that object’s
movements. Similarly, object tracking algorithms enable systems to precisely estimate
the position of an object and follow it over time. Object detection and tracking have
numerous applications, including security surveillance for detecting movement in a
restricted area or recognizing faces in a crowd. Tracking algorithms can be used to
track the movement of cars in real time, which are important for developing autono-
mous vehicles. Additionally, object detection and tracking are utilized in entertain-
ment applications, such as augmented reality, gaming, and virtual reality, where
users need to interact with virtual objects in the environment. Overall, object detec-
tion and tracking are two important computer vision tasks that have led to countless
advances in robotics, automotive engineering, security surveillance, and entertain-
ment. As the development of object detection and tracking algorithms improve, their
applications will only continue to grow.
Image classification and segmentation are important areas of computer vision since
they are used to further understand the content of images. Along with detection,
these methods form the cornerstones of machine vision, which in turn serves as the
foundation for ML. Image classification is the process of assigning a label to an image,
based on the content of the image. This is done by either applying a set of predefined
38 Shaik Mohammad Rafee et al.
The advent of image generation and style transfer technology has revolutionized the
way we interact with digital media. Image generation refers to the process of creating
a unique, synthetic image from computer-generated data. Style transfer is a process
in which the “style” – or visual elements – of one image is applied to the content of
another image, creating a unique effect. In the past, generating a synthetic image
would have required a skilled artist or graphic designer. Through the use of computer
algorithms, however, an AI system is able to do the work of many people – in the
fraction of the time. This is because, unlike humans, computers are able to use a large
number of preexisting inputs to generate new images. Style transfer algorithms are
also able to create new and unique images from existing ones. One common style
transfer algorithm is called “style transfer networks.” These algorithms take a preex-
isting image or “style” and blend it with the content of another image to create a new,
hybrid image. For example, a style transfer network may take the style of a Monet
painting and apply it to a photograph of a castle, creating a completely new image
with both elements from the original two. Image generation and style transfer tech-
nology have a variety of applications, from marketing and advertising to virtual real-
ity and gaming. By adding variety and creativity to otherwise mundane images, this
technology can help bring ideas and products to life, enhancing visual storytelling
and providing an immersive experience for viewers. It can also be used to restore an-
tiquated art, allowing it to be digitally transformed and preserved for future genera-
tions to enjoy. The implications of image generation and style transfer technology are
numerous and profound. By leveraging the power of computer algorithms, designers
and visual artists can focus more of their efforts on artistic expression and creativity.
In turn, these technologies can help us explore different cultures, tell unique stories,
and reimagine the world around us.
Video analysis and understanding focuses on techniques to analyze and gain insight
from different types of videos. It covers a variety of methods, both qualitative and
quantitative, including visual analysis, facial recognition, sound analysis, and object
detection. The analysis begins with a study of the video. This can include aspects such
as frame rate, brightness, contrast, focus, and composition. Other aspects of the video
that can be analyzed are sound, color, and motion. Audio can be analyzed to deter-
mine if there are any repeating patterns or know the types of sounds being used.
Color analysis can help to identify objects and people in scenes as well as provide in-
formation about the time and place being filmed. Motion analysis can provide insight
into how people or objects are moving within the scene. Once the video is analyzed,
the data can be used to build visual models. This involves creating a 3D model of the
40 Shaik Mohammad Rafee et al.
scene to show how people and objects interact in the environment. The models can be
used to create animations that better explain the data. The models can also be used to
track people and objects or to track the changes in the environment over time. The
video analysis and understanding can also help with facial recognition and object de-
tection. Facial recognition can be used to identify a person from a certain group of
people or to verify the identity of a person. Object detection can be used to track ob-
jects in the scene. This is useful in cases such as facial recognition, where the same
person could have multiple appearances. Video analysis and understanding is an im-
portant tool for researchers and businesses to gain new insight from videos. It can be
used to identify trends, detect anomalies, and guide decision-making. It also helps to
protect against fraud, since videos can be used to detect unusual behaviors or activi-
ties. By using the techniques of video analysis and understanding, researchers and
businesses can gain new insights from videos and create better strategies for the
future.
The integration of AI into robotics and automation is revolutionizing the production in-
dustry, and will be key to the future of many industries. AI-driven technologies, such as
robotics and automation, have already proven to be a valuable asset in many different
sectors. The increased use of AI in robotics and automation allows for a more efficient
and accurate production process. Robots and automated systems have already found a
number of different applications, ranging from assembly lines to hazardous waste man-
agement. Autonomous robots are able to quickly and efficiently process and sort, as-
semble, and move parts and components in a controlled manner, drastically increasing
productivity and precision, while simultaneously reducing labor costs. AI-enabled ro-
bots can also carry out a myriad of other tasks through voice commands and have the
ability to learn from their experiences. This learning capability is creating unprece-
dented opportunities for the use of robots and simple automation systems to perform
more complex tasks and to operate in unpredictable and unfamiliar environments. AI-
enabled robots and automation systems also promise to create greater efficiency and
safety at the workplace. In addition to reducing the need for manual labor in hazardous
and difficult-to-access environments, AI-enabled automation also reduces the risk of
human error. Furthermore, systems designed with AI are capable of monitoring their
processes in real time and providing feedback to the system operator, allowing for
greater oversight and accuracy. The introduction of AI in robotics and automation also
enables a new level of customization to the production process. AI-enabled systems
have the ability to recognize patterns, learn from experience, and respond accordingly.
This adaptation and customization makes robots and automation systems more resil-
ient and able to accommodate changes in the output requirements and demands in real
time. The integration of AI into robotics and automation is already having a profound
impact on the production industry. By reducing the dependence on labor-intensive pro-
cesses, AI-enabled systems provide greater efficiency, accuracy, and safety, while simul-
taneously reducing production costs. Through their ability to adapt to changes and
provide real-time feedback, AI-enabled robots and automation systems are also unlock-
ing new opportunities for customizing production processes to meet ever-changing cus-
tomer demands.
Perception and sensing technologies are the foundation of any modern technological
infrastructure, and can be used to improve the efficiency, safety, and effectiveness of
42 Shaik Mohammad Rafee et al.
any system. They allow systems to detect environmental data, monitor activities, in-
teract with users, and identify user behavior. When these technologies are used to
their maximum potential, they can provide immense benefits. The technology works
by capturing data through any number of sensors, whether they are sound, vision, or
motion. This data can be used to create a representation of the reality at hand, which
can then be further analyzed to determine patterns of behavior or identify anomalies.
This allows systems to quickly and accurately identify emerging issues, and respond
appropriately. For example, if a system can detect motion, it may be able to detect
suspicious behavior or the presence of a security threat. On the other hand, cameras
can be used to quickly scan an area to ensure that the environment is safe and secure.
This data can also be used to identify unusual events or behaviors, alerting users to
potential issues before they become serious [32]. Perception and sensing technologies
can also be used to engage with users and influence their behavior. By providing per-
sonalized feedback and interactive experiences, companies can influence consumer
behavior in ways that can have a significant impact on their bottom line. By monitor-
ing user behavior, companies can detect user preferences and habits, and tailor their
products and services accordingly. Perception and sensing technologies are quickly
becoming the norm in modern technology. They are being integrated into more and
more systems, with the potential to significantly enhance security, safety, and effi-
ciency. As these technologies continue to develop, they will likely become a corner-
stone of many industries and applications.
Motion planning and control is an important aspect of robotic engineering and design.
These two fields are concerned with how robots move and interact with their envi-
ronment, and how to control their movements. Motion plans are created in order to
tell a robot how to maneuver in a given environment, with the least amount of time
and energy. This involves the use of various algorithms, sensors, and feedback from
the environment. Control is then applied to ensure proper execution of the motion
plans. Motion planning is a complex process that relies on complex algorithms. It has
evolved over time, from the development of models and simulations to the use of ML
and AI. In the past, motion planning was done manually with trial and error. How-
ever, researchers have developed more efficient algorithmic approaches that enable
robots to craft plans based on observational data from their environment. Control is
the mechanism for executing motion plans. It incorporates feedback from sensors
and the environment to ensure that the robot behaves as desired. This approach can
be used to process information from different sources to make decisions on how to
move in order to reach a given goal. Motion planning and control are closely related
to many fields of engineering, from robotics to autonomous driving. A combination of
motion planning and control enables robots to maneuver safely and reliably in differ-
2 AI technologies, tools, and industrial use cases 43
ent environments while minimizing the energy needed for movement. As the technol-
ogy continues to develop, motion planning and control will enable robots to navigate
through more complex and unpredictable environments with greater accuracy and
speed.
Human-computer interaction is a rapidly growing field that explores how humans in-
teract with and use computers, robots, and other types of technology. It seeks to un-
derstand how humans process, comprehend, and use the systems to interact and
collaborate with computers, robots, and other technological devices. With robots
being increasingly used in many aspects of daily life, it is becoming increasingly im-
portant to understand the interaction between humans and robots. Robots can bring
a lot of benefits to humans, from taking on labor-intensive tasks to providing physical
assistance in industrial and medical settings. This is why there is a growing interest in
studying the human-robot interaction, as the prospect of robots and humans working
in tandem is an exciting discussion in AI research. As robots become increasingly so-
phisticated, there are added complexities to consider when it comes to studying the
human-robot interaction. For instance, the way in which a robot responds to a given
command will depend on the programming, data, and even the way in which the
human responds. It is thought that in the long-term, a robot can eventually possess
enough complexity to think for itself and make decisions independently [33]. The im-
plications of this for human-robot interaction are both interesting and challenging.
There are already some ethical issues surrounding the idea that a robot could take
decisions on its own, which could have an impact on humans. The importance of un-
derstanding how humans interact with robots extends further than just providing ro-
bots with human-level intelligence. By understanding how humans interact with
robots, we can develop better ways in which robots and people can be more produc-
tive and connected. With ongoing research, innovation, and progress, human-robot
interaction could be a completely new field of collaboration and understanding in the
future. The human-robot interaction is an important field of study that seeks to un-
derstand how humans interact with robots, and intends to provide a bridge between
humans and robots that will enable the two to be more productive and connected. As
robots become increasingly sophisticated, there are a variety of factors to consider,
which makes it a fascinating and intriguing area of research.
Industrial automation is the use of technology, systems, and software to control and
monitor industrial processes, machines, and equipment. It has enabled greater pro-
44 Shaik Mohammad Rafee et al.
ductivity and efficiency in industrial operations, with increased safety, reliability, and
accuracy. Industrial automation has revolutionized many industries, from automotive
manufacturing to food production. Industrial automation is most commonly seen in
manufacturing processes, where tasks are repeated and automated using machines
and robots [34]. Automation makes it possible to increase the speed and accuracy of
the production process. It also improves the quality of the product. Industrial automa-
tion also helps in eliminating human error and reducing costs. Industrial automation
can be used in a variety of industries and fields. Some of the most common use cases
are found in the manufacturing industry, where tasks like quality control and inspec-
tion are often automated. Automation also makes it possible to create complex prod-
ucts and reduce the number of workers on the assembly line. In the healthcare sector,
industrial automation is used to streamline processes and tasks, such as analyzing
and interpreting patient data. In the agriculture industry, robotic operations are used
to automate the planting, nurturing, and harvesting of crops. In the transportation
industry, automated trucks are used to deliver goods as well as improve safety and
reduce emissions. Industrial automation has ushered in a new era of efficiency and
productivity in many industries. It helps to reduce costs, increase safety and accuracy,
and improve product quality. With the continued development of technology and soft-
ware, there will likely be new use cases of industrial automation in the future.
tasks, such as surveying large geographic areas or inspecting construction sites from a
bird’s eye view. Although these technologies have practical applications, there are still a
variety of safety and ethical concerns associated with them. Autonomous vehicles and
drones need to be thoroughly tested to ensure they are safe for use on public roads and
operated according to legal restrictions. Additionally, there is a fear that autonomous
vehicles and drones could increase the risk of surveillance, making it easier for compa-
nies and governments to track the movements of citizens. Finally, there are concerns
related to the potential for autonomous vehicles and drones to be hacked and used for
malicious purposes. Autonomous vehicles and drones are reshaping the landscape of
transportation and have the potential to revolutionize industries [36]. In spite of poten-
tial drawbacks, the advantages of autonomous vehicles and drones suggest that they
will continue to play a crucial role in the future of transportation.
The role of AI in finance and banking has seen a rapid growth in recent years. AI is a
technology that enables machines to think, learn, and act like humans. It is being im-
plemented in financial services, across various areas including banking, investments,
insurance, and credit services. AI can help financial services companies to identify
patterns in the customer’s data and provide better assistance to customers. For in-
stance, AI can detect changes in customer spending and alert the customers for possi-
ble frauds. AI can also help in analyzing customer behavior, predicting customer’s
future needs and suggesting appropriate products or services. AI can be used to auto-
2 AI technologies, tools, and industrial use cases 47
mate mundane tasks in financial services, like payments, document handling, and
compliance with regulatory requirements. It can be used to detect anomalies in cus-
tomer’s transactional activities and stop suspicious activities quickly. AI can also be
used to detect frauds, mitigate financial risks, and prevent money laundering activi-
ties. AI can also be used in financial services to create customer profiles for better
segmentation and target marketing. AI-based customer acquisition strategies can also
be used to identify potential customers and reach out to them. AI can be used to offer
personalized advice to individual customers and suggest suitable products that may
suit their needs. AI can also be used to provide automated and optimized customer
service. AI has become essential in the finance and banking industry. It provides com-
panies the power to identify patterns in data, automate mundane tasks, detect frauds,
prevent money laundering activities, create customer profiles, and offer personalized
advice. AI is playing a significant role in transforming the financial services industry
and aiding its growth.
AI has revolutionized many aspects of how businesses run their operations. It has been
especially noteworthy in the fields of manufacturing and supply chain management, and
the benefits of using AI to manage and optimize manufacturing processes and supply
chain operations are felt across every stage of the production cycle. First and foremost, AI
helps to minimize costs by improving efficiency and quickly detecting errors in production
processes. Automating processes that were done manually in the past, and implementing
automated quality control systems, allow managers to quickly identify inefficiencies in
production and fix them with minimal disruption. For supply chain management, AI
quickly identifies any potential problems in the supply chain process, such as supply short-
ages or production delays, and provides real-time solutions that can be implemented
quickly and easily. AI also helps to increase product quality and customer satisfaction. By
accurately predicting customer demand, AI can help businesses identify which products
are in high demand, and ensure the necessary supplies and manpower are available at
the right time to produce these items. AI can also help to keep track of the accuracy and
quality of products, and alert managers when something is not up to standard. AI helps to
reduce the environmental footprint of manufacturing. Data-driven production and AI-
assisted supply chain management allow for optimized production processes that mini-
mize wastage. This is done by optimizing for things like material utilization and reducing
waste in production, as well as optimizing shipping and delivery to minimize fuel costs. AI
has had a transformative effect on the manufacturing and supply chain management pro-
cesses. By reducing costs, increasing product accuracy, and reducing environmental
waste, AI helps businesses to be both cost-effective and sustainable. With AI, businesses
can streamline their operations, improve quality, and increase customer satisfaction.
48 Shaik Mohammad Rafee et al.
The AI technology revolution is rapidly transforming the customer service field as more
and more companies are looking to incorporate AI into their service operations. AI-
powered chatbots are especially becoming popular as they automate various customer
service processes, starting from simple tasks to complex ones. Chat bots are able to un-
derstand customer inquiries, respond effectively, and even simulate a human-like con-
versation with customers, when necessary. They are less expensive than human labor
and can provide faster response times. Furthermore, they have the potential to ensure
customers are getting the best possible customer service experience, since they are able
to process multiple requests simultaneously without having to remember individual in-
formation. AI-powered chatbots can also be used to quickly respond to common queries
and help customers get answers to their questions quickly. For example, if a customer
wants to avail a service or ask for assistance, a chatbot can provide answers to the cus-
tomer quickly and accurately. They can also suggest other services that the customer
may want to explore, based on their inquiry. In addition to responding to customer in-
quiries and resolving issues quickly, AI-friendly chatbots can also provide proactive cus-
tomer service. They can remind customers about existing orders or inform them when
new products become available. Chatbots can also be used to anticipate needs and offer
additional services, such as discounts or vouchers. This can further improve the cus-
tomer service experience and boost customer loyalty. AI technology is having a huge
impact on the customer service field. AI-enabled chatbots are proving to be beneficial
in reducing costs, providing faster service resolution, and providing proactive services
for customers. Companies should continue to explore how this technology can improve
their customer service operations.
The use of AI in energy and utilities is transforming how companies in the sector
manage resources. AI is being used to optimize energy consumption and production,
identify problems quickly, and improve customer service. AI can optimize energy con-
sumption and production by using advanced analytics to discover new ways to save
energy and improve efficiency. It can analyze data from different sources, such as
customer bills and weather forecasts, and suggest cost-effective solutions. AI can iden-
tify variations in energy usage that could potentially indicate waste or an energy leak
and alert the utility company to take action. AI can also predict how usage patterns
may change in the near future, and adjust operations accordingly.
2 AI technologies, tools, and industrial use cases 49
People everywhere are facing the challenge of feeding the world’s growing population
while using fewer resources. To help meet this challenge, many people have started to
look to AI for assistance in agriculture. AI is an ever-evolving field that has revolution-
ized various industries such as healthcare, finance, and education. Now, AI is begin-
ning to take on a major role in agriculture and farming. AI has the potential to
monitor environments and crops better than manual labor. AI applications now allow
farmers to accurately recognize weeds and pests, count individual crops, and identify
soil health issues. Farmers are using AI to create a “digital twin” of their farm – an
exact 3D representation of the farm and its inputs, which can be used to make better
informed decisions and optimize operations. With AI, fertilizers, irrigation, and pesti-
cides can be applied more precisely and in smaller doses [39]. AI can help farmers
maximize the use of their resources by enabling them to identify the optimal amounts
and combinations of inputs for any given crop. AI can also streamline the farming
process by automating repetitive activities such as seeding and harvesting. AI is also
being used for predictive analysis. AI can analyze a variety of variables such as
weather patterns, crop production, soil data, and market prices. With this data, farm-
ers can make more informed decisions on where to invest their resources for maxi-
mum efficiency. However, there are many obstacles that must be overcome for AI to
be adopted widely in agriculture. AI technologies are still expensive and require ex-
perts to ensure the systems provide accurate results. Accessibility to the internet is a
major issue for many rural farmers, a factor which limits the effectiveness of AI tech-
nologies. In spite of the challenges, AI has the potential to revolutionize agriculture. It
has the potential to make farming more efficient, accurate, and resourceful. As the
technology matures, we will see even greater advancements in AI’s role in farming
and agriculture.
References
[1] Patel, P., Ali, M. I., & Sheth, A. (2018). From raw data to smart manufacturing: AI and semantic web
of things for industry 4.0. IEEE Intelligent Systems, 33(4), 79–86.
[2] Augustine, P., Raj, P., & Munirathinam, S. (Eds.). (2022). Enterprise digital transformation:
Technology, tools, and use cases. New York: CRC Press.
[3] Alexopoulos, K., Nikolakis, N., & Chryssolouris, G. (2020). Digital twin-driven supervised machine
learning for the development of artificial intelligence applications in manufacturing. International
Journal of Computer Integrated Manufacturing, 33(5), 429–439.
[4] Heier, J. (2021, July). Design intelligence-taking further steps towards new methods and tools for
designing in the age of AI. In Artificial Intelligence in HCI: Second International Conference, AI-HCI
2021, Held as Part of the 23rd HCI International Conference, HCII 2021, Virtual Event, July 24–29,
2021, Proceedings (pp. 202–215). Cham: Springer International Publishing.
50 Shaik Mohammad Rafee et al.
[5] Whittaker, M., Crawford, K., Dobbe, R., Fried, G., Kaziunas, E., Mathur, V. . . . Schwartz, O. (2018). AI
now report 2018. (pp. 1–62). New York: AI Now Institute at New York University.
[6] Kakani, V., Nguyen, V. H., Kumar, B. P., Kim, H., & Pasupuleti, V. R. (2020). A critical review on
computer vision and artificial intelligence in food industry. Journal of Agriculture and Food
Research, 2 100033.
[7] Bumblauskas, D., Mann, A., Dugan, B., & Rittmer, J. (2020). A blockchain use case in food
distribution: Do you know where your food has been?. International Journal of Information
Management, 52 102008.
[8] Giordani, M., Polese, M., Mezzavilla, M., Rangan, S., & Zorzi, M. (2020). Toward 6G networks: Use
cases and technologies. IEEE Communications Magazine, 58(3), 55–61.
[9] Chui, M., Roberts, R., & Yee, L. (2022). Generative AI is here: How tools like ChatGPT could change
your business. Quantum Black AI by McKinsey, 1–5.
[10] Kuzior, A., Sira, M., & Brozek, P. (2022). Using blockchain and artificial intelligence in energy
management as a tool to achieve energy efficiency. Virtual Economics, 5(3), 69–90.
[11] Chou, J. S., Tai, Y., & Chang, L. J. (2010). Predicting the development cost of TFT-LCD manufacturing
equipment with artificial intelligence models. International Journal of Production Economics, 128(1),
339–350.
[12] Schlögl, S., Postulka, C., Bernsteiner, R., & Ploder, C. (2019). Artificial intelligence tool penetration in
business: Adoption, challenges and fears. In Knowledge Management in Organizations: 14th
International Conference, KMO 2019, Zamora, Spain, July 15–18, 2019, Proceedings 14 (pp. 259–270).
Spain: Springer International Publishing.
[13] Shaw, J., Rudzicz, F., Jamieson, T., & Goldfarb, A. (2019). Artificial intelligence and the implementation
challenge. Journal of Medical Internet Research, 21(7), e13659.
[14] Praghash, K., Arshath Raja, R., Chidambaram, S., & Shreecharan, D. (2023, May). Hyperspectral
image classification using Denoised Stacked Auto Encoder-Based Restricted Boltzmann Machine
Classifier. In Ajith Abraham,Tzung-Pei Hong, Ketan Kotecha, Kun Ma, Pooja Manghirmalani Mishra,
& Niketa Gandhi (Eds.). Hybrid Intelligent Systems: 22nd International Conference on Hybrid
Intelligent Systems (HIS 2022), December 13–15, 2022 (pp. 213–221). Cham: Springer Nature
Switzerland.
[15] Bello, O., Holzmann, J., Yaqoob, T., & Teodoriu, C. (2015). Application of artificial intelligence
methods in drilling system design and operations: A review of the state of the art. Journal of
Artificial Intelligence and Soft Computing Research, 5(2), 121–139.
[16] Lee, D., & Yoon, S. N. (2021). Application of artificial intelligence-based technologies in the
healthcare industry: Opportunities and challenges. International Journal of Environmental Research
and Public Health, 18(1), 271.
[17] Davenport, T. H. (2018). From analytics to artificial intelligence. Journal of Business Analytics, 1(2),
73–80.
[18] Sivapriya, T. R., & Saravanan, V. (2011). Automatic brain MRI mining using Support Vector Machine
and Decision tree. CiiT International Journal of Artificial Intelligent Systems and Machine Learning,
3(2), 109–116.
[19] Kohli, M., Alkasab, T., Wang, K., Heilbrun, M. E., Flanders, A. E., Dreyer, K., & Kahn, C. E., Jr (2019).
Bending the artificial intelligence curve for radiology: Informatics tools from ACR and RSNA. Journal
of the American College of Radiology, 16(10), 1464–1470.
[20] Peres, R. S., Jia, X., Lee, J., Sun, K., Colombo, A. W., & Barata, J. (2020). Industrial artificial intelligence
in industry 4.0-systematic review, challenges and outlook. IEEE Access, 8 220121–220139.
[21] Bhende, M., Thakare, A., Pant, B., Singhal, P., Shinde, S., & Saravanan, V. (2022). Deep learning-
based real-time discriminate correlation analysis for breast cancer detection. BioMed Research
International, 2022, 1–12.
2 AI technologies, tools, and industrial use cases 51
[22] Ren, L., Zhang, L., Wang, L., Tao, F., & Chai, X. (2017). Cloud manufacturing: Key characteristics and
applications. International Journal of Computer Integrated Manufacturing, 30(6), 501–515.
[23] Romanovs, A., Pichkalov, I., Sabanovic, E., & Skirelis, J. (2019, April). Industry 4.0: Methodologies,
tools and applications. In 2019 Open Conference of Electrical, Electronic and Information Sciences
(eStream) (pp. 1–4). IEEE.
[24] Praghash, K., Peter, G., Stonier, A. A., & Priya, R. D. (2023, May). Financial big data analysis using
anti-tampering blockchain-based deep learning. In Hybrid Intelligent Systems: 22nd International
Conference on Hybrid Intelligent Systems (HIS 2022), December 13–15, 2022 (pp. 1031–1040). Cham:
Springer Nature Switzerland.
[25] Hansen, E. B., & Bøgh, S. (2021). Artificial intelligence and internet of things in small and medium-
sized enterprises: A survey. Journal of Manufacturing Systems, 58 362–372.
[26] Rabah, K. (2018). Convergence of AI, IoT, big data and blockchain: A review. The Lake Institute
Journal, 1(1), 1–18.
[27] Sircar, A., Yadav, K., Rayavarapu, K., Bist, N., & Oza, H. (2021). Application of machine learning and
artificial intelligence in oil and gas industry. Petroleum Research, 6(4), 379–391.
[28] Silvia Priscila, S., Sathish Kumar, C., Manikandan, R., Yuvaraj, N., & Ramkumar, M. (2022, February).
Interactive artificial neural network model for UX design. In International Conference on
Computing, Communication, Electrical and Biomedical Systems (pp. 277–284). Cham: Springer
International Publishing.
[29] Sayed-Mouchaweh, M. (Ed.). (2020). Artificial intelligence techniques for a scalable energy
transition: Advanced methods, digital technologies, decision support tools, and applications. India:
Springer Nature.
[30] Avinash Khatri, K. C., Shah, K. B., Logeshwaran, J., & Shrestha, A. (2023). Genetic algorithm based
techno-economic optimization of an isolated hybrid energy system. ICTACT Journal on
Microelectronics, 8(4), 1447–1450.
[31] Manikandan, R., Sara, S. B. V., Yuvaraj, N., Chaturvedi, A., Priscila, S. S., & Ramkumar, M. (2022, May).
Sequential pattern mining on chemical bonding database in the bioinformatics field. In AIP
Conference Proceedings (Vol. 2393, No. 1, p. 020050). AIP Publishing LLC.
[32] Ramesh, G., Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). Prediction of energy production level
in large PV Plants through AUTO-Encoder Based Neural-Network (AUTO-NN) with restricted
Boltzmann feature extraction. Future Internet, 15(2), 46.
[33] Kumar, A., Yadav, S. P., & Mittal, S. (2020). DAI for document retrieval. In : Satya Prakash Yadav,
Dharmendra Prasad Mahato & Nguyen Thi Dieu Linh (Eds.).Distributed Artificial Intelligence
(pp. 251–262). Boca Raton: CRC Press.
[34] Tang, A., Tam, R., Cadrin-Chênevert, A., Guest, W., Chong, J., & Barfett, J., Canadian Association of
Radiologists (CAR) Artificial Intelligence Working Group (2018). Canadian Association of Radiologists
white paper on artificial intelligence in radiology. Canadian Association of Radiologists Journal,
69(2), 120–135.
[35] Al-Turjman, F., Yadav, S. P., Kumar, M., Yadav, V., & Stephan, T. (Eds.). (2022). Transforming
management with AI, Big-Data, and IoT. India: Springer.
[36] Parunak, H. V. D. (1996). Applications of distributed artificial intelligence in industry. Foundations of
Distributed Artificial Intelligence, 2(1), 18.
[37] Alippi, C., Ferrero, A., & Piuri, V. (1998). Artificial intelligence for instruments and measurement
applications. IEEE Instrumentation and Measurement Magazine, 1(2), 9–17.
[38] Marr, B. (2019). Artificial intelligence in practice: How 50 successful companies used AI and machine
learning to solve problems. USA: John Wiley & Sons.
[39] Bravo, C., Saputelli, L., Rivas, F., Pérez, A. G., Nikolaou, M., Zangl, G. . . . Nunez, G. (2014). State of
the art of artificial intelligence and predictive analytics in the E&P industry: A technology survey. SPE
Journal, 19(04), 547–563.
G. Venkataramana Sagar, S. Ambareesh, P. John Augustine, R. Gayathri
3 Classification and regression algorithms
Abstract: Machine learning algorithms for predicting or categorizing data include
classification and regression techniques. Regression algorithms are used to forecast a
continuous numerical value, such as the anticipated value of a stock price. In contrast,
classification algorithms predict a discrete label, such as whether something is a cat
or a dog. The accuracy of the algorithm’s predictions and classifications determines
how effective a prediction or classification is. Decision trees, support vector machines,
k-nearest neighbors, and naive Bayes are examples of standard classification techni-
ques. Regression methods often used include generalized, polynomial, and linear
regression.
https://doi.org/10.1515/9783111323749-003
54 G. Venkataramana Sagar et al.
To precisely predict the intended learning output, the supervised learning approach
needs a “training dataset.” This training dataset consists of labeled samples, including
the desired result and the input data. The labeled data is used to train the model on
how to predict the intended outcome from new data. Following the supervised learn-
ing procedure’s completion, the model should be able to predict and categorize new
data points correctly. This makes it possible to test the model using inputs that it has
not previously been trained on. For computer scientists and data scientists, super-
vised learning is a potent tool that enables automatic predictions from new data or
data the model has never seen before. The model can develop the ability to anticipate
the desired outcomes more precisely as it continues to gather more data and is
trained with additional labels.
ML projects need training and testing datasets because they give the model the knowl-
edge and experience to make wise decisions and predictions. The testing dataset is
used to assess the model’s correctness and ascertain how well it will make a predic-
tion or decision on previously unknown data. The training dataset is used to create
and train the model. Training datasets are gathered via prior experience and are used
to instruct the model on how to correctly find patterns, correlations, and similarities
among various parameters [4]. One or more associated factors, such as a client’s
name, age, marital status, and spending patterns, may be included in a training data-
set that the model can use to determine which good or service will most likely meet
the consumer’s demands. Testing datasets differ from others because they contain in-
formation the model has never seen before. This information evaluates the model’s
ability to conclude, make predictions, or compare data that it has never seen before.
Testing datasets are also used to ensure the model is suitably extended to fit many
circumstances rather than overfitting the training data. Training and testing datasets
are necessary to build and assess an ML model. Using both datasets, the model can
learn how to make accurate predictions and decisions based on experience and data
it has never seen. ML projects need training and testing datasets because they give
the model the knowledge and expertise to make wise decisions and predictions.
The testing dataset is used to assess the model’s correctness and ascertain how
well it will make a prediction or decision on previously unknown data. The training
dataset is used to create and train the model. Training datasets are gathered via prior
experience and are used to instruct the model on how to correctly find patterns, cor-
relations, and similarities among various parameters [4]. One or more associated fac-
tors, such as a client’s name, age, marital status, and spending patterns, may be
included in a training dataset that the model can use to determine which good or ser-
56 G. Venkataramana Sagar et al.
vice will most likely meet the consumer’s demands. Testing datasets differ from
others because they contain information the model has never seen before. This infor-
mation evaluates the model’s ability to conclude, make predictions, or compare data
that it has never seen before. Testing datasets are also used to ensure the model is
suitably extended to fit many circumstances rather than overfitting the training data.
Training and testing datasets are necessary to build and assess an ML model. Using
both datasets, the model can learn how to make accurate predictions and decisions
based on experience and data it has never seen.
Metrics for evaluation are used to gauge a model’s efficacy. This is crucial when con-
trasting various models to determine which of them fits a particular issue the best.
Two of the most popular ML approaches are classification and regression, and there
are multiple metrics to gauge their effectiveness. Regression measures include mean
absolute error, mean squared error, and R2 score, whereas classification metrics in-
clude accuracy, precision, recall, and F1 score. The most used classification metric is
accuracy, which expresses how effectively a model predicts as a percentage of all in-
stances [5]. Higher scores indicate more accurate predictions. Precision is calculated
as the ratio of genuine positives to all anticipated positives. Higher scores in recall
indicate fewer false negatives since it gauges the proportion of real positives to actual
positives. The F1 score combines precision and recall, and the highest scores show the
best coordination of these two metrics. The average discrepancy between the antici-
pated and actual values is the mean absolute error in regression. The mean squared
error measures the average of the squares of the differences between the predicted
and actual values. The percentage of data variation that the model successfully ex-
plains is the R2 score. This gauges how well the model can account for any observed
data changes. Accuracy, precision, recall, and F1 score are assessment metrics for clas-
sification; mean absolute error, mean squared error, and R2 score are evaluation
measures for regression. These measures can be used to compare the performance of
many ML models and guarantee that the best model is selected for a given task.
and categorizes the likelihood of a target outcome. The observed data are mapped
onto a range of values between 0 and 1 using a logistic function. As it can forecast the
likelihood that an event or consequence will occur, logistic regression is a valuable
tool in many different domains [6]. A simple strategy that yields quick and precise
results is logistic regression. It may be applied to various tasks, including predicting
stock prices, predicting customer turnover, and classifying data into multiple groups.
It is also very adaptable because it may be applied to both linear and nonlinear mod-
els. Precision, recall, accuracy, and F1 scores are frequently used to evaluate logistic
regression models. These metrics gauge how well the model predicts the desired re-
sult or class. They also aid in locating potential biases and error-producing factors in
the model. A versatile predictive modeling method that may be employed in many
different disciplines is logistic regression. This flexible approach can handle both lin-
ear and nonlinear models. It is simple to use and yields quick, precise results. Addi-
tionally, its performance is frequently assessed using F1 scores, precision, recall, and
accuracy measures.
A type of statistical analysis called binary logistic regression is used to forecast the
likelihood that a dichotomous (two-category) event will have a particular outcome in
the future. This type of regression analysis uses continuous independent variables
and a categorical dependent variable. In light of the knowledge offered by the ob-
served data, it is used to calculate the likelihood that a specific occurrence will occur.
Binary logistic regression models are used for outcomes with just two possible values,
such as true or false, male or female, pass or fail, yes or no, on or off, etc. Binary logis-
tic regression is a powerful technique to model binary response data and comprehend
correlations between independent factors and dependent variables. It can be used to
calculate the likelihood that a specific outcome will occur, making it valuable for fore-
casting future behavior. It can also pinpoint crucial elements that affect how well a
forecast turns out. To comprehend how different circumstances and events affect out-
comes or results, it is frequently used in medical, psychological, marketing, and other
disciplines [7]. The relationship between the dependent variable and one or more in-
dependent factors that affect the result is explained using the binary logistic regres-
sion model. This is accomplished by evaluating the impact of each independent
variable on the likelihood of the outcome. In binary logistic regression, the indepen-
dent variables are assumed to be unrelated and have a linear connection with the
event’s probability. Use logistic regression to comprehend how several independent
variables affect a decision’s outcome. It can be used, for instance, to investigate how
various client characteristics may affect their propensity to sign up for a service. It
can also assess how well a particular marketing effort attracted new clients. Binary
logistic regression can be used in the medical industry to evaluate how risk factors
58 G. Venkataramana Sagar et al.
affect the likelihood of developing disease and dying. Binary logistic regression is a
valuable technique for modeling, forecasting, and comprehending the possibility that
specific events will occur in the future. It is a useful tool for many different fields and
has a wide range of applications.
Multinomial logistic regression is a form of ML technique that works well for categori-
cal outcome prediction. A set of explanatory variables, commonly called predictors,
are used to train a logistic regression model to forecast the likelihood of a discrete
result. In contrast to binary logistic regression, which is used to classify data with
only two possible outcomes, multinomial logistic regression can be used to classify
data with more than two possible outcomes. Multinomial logistic regression can ana-
lyze data effectively and make precise predictions, such as forecasting a customer’s
response to an offer or placing them in a specific market group. A dataset can be ex-
amined for trends and associations using multinomial logistic regression, which can
then be used to make decisions or develop marketing plans. Compared to other ML
techniques, multinomial logistic regression models are thought to be more robust and
better able to withstand overfitting [8]. This is because it uses the links between the
explanatory variables to forecast the outcome rather than assuming that one variable
is the only significant predictor of a given event. As a result, multinomial logistic re-
gression frequently makes more precise forecasts. Multinomial logistic regression is
typically implemented using software like SAS or R. With the appropriate datasets
and model parameters, this kind of software is intended to support the multinomial
logistic regression process.
Additionally, the model’s outputs may be interpreted using the software, which
helps comprehend the influence that each predictor has on the conclusion. The model
can then be utilized to generate accurate predictions after it has been put into prac-
tice. Multinomial logistic regression is a robust statistical method that can be used to
categorize data with several outcomes precisely. To create accurate predictions, it en-
ables the evaluation of the associations between the explanatory factors and the re-
sult. Additionally, this ML method can be used to find patterns and insights in data,
helping decision-makers.
When one or more predictor variables (X) are about a binary class variable (Y), logis-
tic regression is a statistical approach used to examine the data. It is one of the most
popular predictive modeling approaches in the predictive analytics community. For
classification issues where the goal variable is binary, logistic regression predicts cate-
3 Classification and regression algorithms 59
Decision trees are a helpful tool that customers and business owners can use to un-
derstand their data and make informed decisions. A diagrammatic representation of
the decision-making process, decision trees show each option and its effects. By
quickly understanding the various possibilities for a decision, a firm can receive in-
sight into where and how to move forward, avoiding costly mistakes that might occur
if poor judgment is made. Making decision trees requires the use of input data. This
information could consist of market research, financial data, or customer surveys, de-
pending on the kind of decision that needs to be made. This creates a decision tree,
with each branch denoting a potential outcome of the choice. Following each chapter,
the decision-maker can determine which product is preferable. The nodes represent
the choices that must be made, and the links between the nodes show the potential
courses of action. The tree’s leaves or endpoints characterize the conclusion of the
decision-making process.. The branches are scored or evaluated using the decision-
maker’s stated criteria, and the best outcome is chosen. Decision trees can tackle
problems such as client segmentation, customer service, product testing, and financial
analysis. It is important to note that decision trees should only be used for classifica-
tion problems, not regression problems [11]. The decision tree is a reliable data analy-
sis method that can streamline and simplify difficult decisions. By visually presenting
data and potential courses of action, decision trees help business owners, decision-
makers, and customers quickly discover and choose the appropriate action.
disorganized or ensure it is. Finding the optimal point for categorizing our data into
different classes is the goal of decision trees. Information gain is a metric representing
the anticipated entropy reduction by dividing the data by an attribute. Typically, the
point with the most significant information gain is preferred for the split. It can em-
ploy a calculation like the Shannon entropy, which considers the likelihood that each
sample will belong to a specific class, to determine the information gain. The degree
of uncertainty in the data increases with the Shannon entropy. The knowledge gain is
higher, the less uncertain the situation. It also considers other aspects when selecting
which attribute to choose for a split. In general, it favors details with a bigger sample
size or more different values, enabling us to divide the data more effectively. Al-
though various factors may affect the split, the computed information gain deter-
mines the preferable partition [12]. The feature that best distinguishes data samples
from one another can be chosen using the splitting criteria of decision trees. We can
locate the part with the most significant gain by computing the Shannon entropy,
which allows us to quantify the information gain of each property. When choosing
which feature splits the data most effectively, further factors may also be considered.
Decision trees are one of the most widely used ML techniques for supervised classifi-
cation tasks. They are regarded as one of the most precise and trustworthy prediction
models. A decision tree is the most effective solution for classification and regression
issues. A decision tree’s principal function is to segment a dataset into more digestible
chunks so that predictions can be made. This is accomplished by managing the data-
set’s categorical and numerical properties.
In contrast to numerical features – variables with continuous values like height,
weight, and salary – categorical features are variables with discrete values like gen-
der, race, and age. The decision tree uses a divide-and-conquer strategy to classify the
data when dealing with flat features [13]. According to the values of the categorical
features, it will divide the dataset into subgroups. The decision tree then gives each of
the subgroups a label or class. A decision tree will establish a split point at a specific
value to divide the data into various branches while handling numerical features. The
facts on the different components are then once more labeled. The entropy or infor-
mation gain associated with each characteristic determines this approach’s split
points and labels. The decision tree is likelier to incorporate a feature in its prediction
when the entropy is higher. The decision tree will find the title or class that appears
the most frequently among the branches to make a prediction, and it will then assign
that label or type to the data. This is referred to as the majority voting strategy. When
it comes to handling both category and numerical variables, decision trees are very
effective. In addition, they are a preferred option among other ML algorithms due to
their simplicity and precision.
62 G. Venkataramana Sagar et al.
Due to their better accuracy and resilience, ensemble methods – which incorporate
numerous algorithms – have emerged as one of the most widely used methods for
supervised ML. Gradient boosting and random forests are the most commonly used
ensemble techniques. An ensemble technique called random forests creates numer-
ous decision trees to generate predictions. Each decision tree employs a unique mix
of characteristics and is trained using a different subset of the training data. As a re-
sult, separate decision trees are created, which are subsequently joined using a simple
majority voting strategy. With the help of this averaging strategy, the model’s bias
and variance are reduced, significantly improving its accuracy and robustness. An en-
semble technique called gradient boosting joins several “weak” learners to create a
“strong” learner. Different subsets of data are used to train each weak learner’s deci-
sion tree; however, unlike random forests, the vulnerable learners are not indepen-
dent. The learners that follow are taught to concentrate on the data points the
previous learner struggled to classify accurately. The performance of the weaker
learners determines the weights for the weighted average strategy used to mix these
learners. This method produces a model with much higher accuracy and robustness
than a single decision tree. Gradient boosting and random forests are effective ensem-
ble techniques that can boost precision and robustness. Although each strategy has
pros and cons, they are both successful solutions for supervised learning issues.
The linear SVM is a supervised ML model that recognizes and separates data points in
feature space using several methods. This algorithm creates a hyperplane or collec-
tion of hyperplanes in a high- or infinite-dimensional room that can divide various
object types. The algorithm can make a decision boundary that will assign classes to
the input objects by identifying a group of hyperplanes that collectively divide the
multiple categories of things. To maximize the distance between the two data types,
the linear SVM algorithm first builds two parallel hyperplanes of maximum width,
known as a decision surface. A unique optimization method known as the quadratic
programming method, which optimizes an expression containing the geometric dis-
tance between the two classes of data, is used to find the best hyperplane (maximizing
the width between the two data classes). The program will employ a margin measure-
ment after creating the ideal hyperplane. The margin is the shortest distance between
two data points, one from each of the two classes, in a linear SVM. The algorithm will
determine the margin using the two locations closest to the hyperplane [15]. In a wide
range of classification tasks, it has been discovered that the linear SVM method per-
forms exceptionally well. It is now among the most used algorithms for handling su-
pervised learning issues. The algorithm is well-optimized and has been proven to
scale very well with massive datasets due to its speed and complete accuracy. The lin-
ear SVM algorithm is helpful for many applications since it is resilient against outliers
and can handle high-dimensional datasets.
A potent method for classification jobs, particularly for big and complicated data sets,
is the nonlinear SVM with kernel functions. Algorithms for ML make extensive use of
it. A hyperplane that optimizes the margin or separation between the two classes is
used to map a data set into a higher dimensional space using a kernel function to sep-
arate the points. These models are also called “kernel SVMs” due to their nonlinear
nature. The polynomial kernel, radial basis function (RBF) kernels, and sigmoid ker-
nels are SVM’s most widely used kernel functions. Each operation on specific data
sets outperforms linear kernels like the linear SVM. The polynomial kernel is fre-
quently used when dealing with data sets that cannot be separated linearly, such as
the XOR problem. It converts the provided data points into a higher dimensional
space and divides the ends using a polynomial [16]. When the data cannot be split
linearly, the RBF kernel is used. It is used for cases where the relationship between
variables can be complex and is more flexible than a polynomial kernel. The RBF ker-
nel can be used for various data types and supports different decision boundaries.
When the data cannot be linearly separated, and the output can have a value of 0 or
64 G. Venkataramana Sagar et al.
1, the sigmoid kernel is used. It is used for issues like text classification, where giving
the probability to particular classes is necessary. The sophisticated tools known as
nonlinear SVMs with kernel functions can be used to categorize large, complex data
sets. They are very helpful when handling massive data sets. The complexity of the
data and the task at hand will determine the type of kernel employed.
Since its emergence in the 1990s, SVMs have gained popularity as a tool for various ap-
plications. SVMs are supervised algorithms that categorize data by locating the hyper-
plane that most accurately distinguishes the two classes. Even when the data contains
outliers or noise, it performs well with small datasets and has good accuracy. Regres-
sion, classification, novelty detection, and clustering are just a few of the tasks in which
SVMs can be used. They are highly adaptable and may be used with organized and un-
structured data. Image classification, text classification, object identification, facial rec-
ognition, protein folding, robotic navigation, and handwriting recognition are some of
the applications for SVMs that are most frequently used.
SVMs are frequently used in the classification of images. The SVM model can cor-
rectly categorize new photos of the same items after being trained on a large batch of
labeled photographs of those same objects. Complex traits that conventional algorith-
mic methods might be unable to identify can be found using the SVM model.
3 Classification and regression algorithms 65
These are only a few of the numerous uses and applications for SVMs. SVMs are a
valuable tool for any data scientist or ML engineer because of their adaptability and
capacity to classify data effectively even when it contains noise and outliers.
where P(A|B) is the probability of event A occurring given that event B has already
occurred. P(B|A) is the probability of event B occurring given that event A has already
occurred. P(A) is the probability of event A occurring, independent of any other
events. Finally, P(B) is the probability of event B occurring, independent of any other
events.
For example, let us say we want to use the Bayes’ theorem to calculate the proba-
bility of someone having the flu, given that they have a fever. We could express this
problem in the following equation:
Here, P(flu|fever) is the probability of having the flu given that the person has a
fever. P(fever|flu) is the probability of having a fever, given that the person al-
ready has the flu. Without considering any other influencing circumstances, P(flu) and
P(fever) represent the likelihood of getting the flu and a fever, respectively. Given the
existence of a fever, we may determine the possibility of having the flu by plugging the
actual probabilities of each event into the equation. This allows us to revise our assess-
ment of an event’s likelihood when new information comes to light. As this example
shows, the Bayes’ theorem is a helpful resource for performing probabilistic computa-
tions. Its fundamental ideas are simple, and they may be applied to make decisions
about potential outcomes in various settings.
An ML method called the naive Bayes assumption forecasts potential outcomes based
on a group of connected variables. This presumption, based on the Bayes’ theorem,
asserts that all of the model’s linked variables are independently correlated [19]. This
means that the value of a single variable will not impact the results of any other vari-
ables in the model. This supposition makes computation easier since it eliminates the
need to study the intricate relationships between the variables and enables the model
to generate predictions much more quickly. This assumption simplifies the calculating
process, but it can be wrong if the data are not correctly analyzed; remembering is
3 Classification and regression algorithms 67
crucial. This assumption does not consider potential correlations between the varia-
bles, which means it might not accurately reflect their actual relationship. Under-
standing the connections between the dataset’s variables and using them to create a
trustworthy algorithm is crucial for successfully using the Naive Bayes assumption.
The model must be carefully examined and tested to ensure accuracy, like with any
ML technique. Additionally, it is crucial to verify that the data meet the assumptions
established in the model before relying on the results produced by this assumption
because the independent nature of this assumption depends on the underlying data.
The probabilistic classification algorithm, Gaussian naive Bayes (GNB), is also called a
linear classifier. It is a well-known and effective classification technique based on
Bayes’ theorem and employs a probabilistic classification strategy. It is frequently
used in various applications, including spam filtering, text classification, and NLP.
Given a set of input data, GNB calculates the probability of a specific event using sta-
tistics. It presumes that each characteristic has a normal distribution and that all fea-
tures are independent. Due to its resistance to outliers, this classification approach is
frequently employed in issues requiring numerous classes and is highly helpful for
handling unbalanced or noisy datasets. A basic GNB classification algorithm has two
parts: first, it determines the likelihood of each type for each feature, and second, it
determines the posterior probability of each class for a given set of input values [20].
It also incorporates prior knowledge to make the model more accurate, such as what
the data type should indicate or what the class labels should represent. The simplicity
and execution speed of the GNB algorithm is among its benefits. It may be put into
use very quickly, and compared to specific other approaches, its prediction time is
quicker.
Additionally, the technique is more generalizable and less prone to overfitting be-
cause it only needs a small number of training samples to determine the weights for
the classes and features. In comparison to other probabilistic classification techni-
ques, it requires less processing power. Due to the assumption that all data would
have a normal distribution, the GNB algorithm has drawbacks when dealing with
non-normal data. The GNB algorithm may also produce erroneous predictions if the
prior data is inaccurate. Every classification algorithm, including GNB, is susceptible
to overfitting, which can harm the algorithm’s effectiveness. An effective and practi-
cal technique for ML and data analysis is the GNB. It is a popular option in many sec-
tors since it can be used to answer a wide range of classification problems accurately.
68 G. Venkataramana Sagar et al.
The multinomial naive Bayes is an ML method used in classifying text and other data.
It is based on the Bayes’ theorem and assumes that each feature contributes equally
to classification. It differs from multivariate naive Bayes in that it relies on an individ-
ual probability distribution for each part rather than a joint probability distribution
like multivariate naive Bayes does. The initial step in the algorithm’s operation is de-
termining the probability of a particular class given the input features. Assuming that
each part has a multinomial distribution, this is accomplished. The class with the
highest chance is then chosen as the anticipated class based on the class-specific prob-
ability distribution. Numerous text classification tasks, including sentiment analysis,
topic modeling, language detection, can be performed using the technique.
Additionally, it aids in biological sequence recognition and medical diagnosis [21].
Multinomial naive Bayes is quick, easy, and accurate compared to other ML algo-
rithms. It is effective with a limited dataset and applies when missing data exists. Ad-
ditionally, it is resistant to the influence of unimportant traits. Also, there are several
drawbacks to multinomial Naive Bayes, such as the assumption that each feature is
independent of the others. In some circumstances, this presumption may be accurate
only sometimes.
Additionally, it ignores the potential significance of feature interactions. The mul-
tinomial naive Bayes algorithm is a potent tool for categorizing data. It can be used
for many things, including language identification, sentiment analysis, and medical
diagnosis. With a limited dataset, it can produce good results quickly and accurately.
It is also reasonably resistant to the influence of unimportant factors.
A robust ML technique called naive Bayes can forecast an outcome given a set of
input variables. It is founded on the Bayes’ Theorem, a method for estimating the like-
lihood of certain events. Naive Bayes is a prominent and highly favored option for
many data science projects because of its wide range of applications and use cases.
– Spam filtering is one application where naive Bayes is frequently used. Based on
previously collected data, Naive Bayes can be used to determine whether a given
email is spam. A spam detector can use specific terms or phrases from the email
and the sender’s information. If you have adequate data, you can use Naive Bayes
to develop an efficient classifier for this issue.
– The classification of texts is another use for naive Bayes. For NLP, categorical la-
bels are frequently necessary for textbooks to be processed intelligently. For in-
stance, you might want to group sentences or text documents into various
categories. Naive Bayes can calculate the odds of each text falling into each class
quickly and accurately.
3 Classification and regression algorithms 69
– Predicting client churn can also be done with naive Bayes. By indicating the dan-
gers, naive Bayes can help organizations identify which clients will most likely
stop using their services. This forecast can be formed based on client demo-
graphic information, transaction histories, and other behavioral patterns. This
aids companies in staying one step ahead and implementing the necessary
adjustments.
Sentiment analysis frequently makes use of naive Bayes. The system can determine
whether a sentence is a positive or negative ken. This is a helpful tool for businesses
to quickly and accurately learn what customers think about their goods or services.
Naive Bayes is an ML technique with many applications that is very adaptable
and adaptive. It can be used for sentiment analysis, text classification, customer
churn prediction, etc. Due to this, it is a standard option for many data science
initiatives.
The K-NN algorithm is a classification system that groups objects according to their
characteristics or qualities. It is a nonparametric approach, which implies that it does
not rely on any presumptions regarding how the data are distributed. Instead, it ex-
amines the nearest data points and applies their labels to those points to classify the
70 G. Venkataramana Sagar et al.
new issue. The foundation of the K-NN method is the idea that related objects should
be grouped. The K-NN technique begins by choosing the “K”’ data points nearest to
the newly discovered data point that needs to be categorized. Afterward, a “weight” is
assigned based on separation of each of those data points from the new issue. The
algorithm uses the weighted average of the labels assigned to the unique point’s K-NN
to forecast its title [23]. Applications for the K-NN algorithm include pattern recogni-
tion, image processing, and data mining. It is well-liked in many applications since it
is a simple and quick method. The K-NN algorithm benefits from not making assump-
tions about the underlying distribution of the data and performs well with noisy data.
The K-NN technique is frequently used in a wide range of applications and can be a
valuable tool for identifying objects.
A classification and regression approach called K-NN is used to find the nearest data
points in a dataset. The algorithm must create a distance metric or metric to measure
the separation between two data points to make decisions based on its data points. A
K-NN method can employ a wide range of distance metrics, including Euclidean, Man-
hattan, Minkowski, Chebyshev, Mahalanobis, Hamming, cosine similarity, and Bray-
Curtis. Each measure has unique qualities and is valuable for specific kinds of data.
– The Euclidean distance metric is the most used distance metric for K-NN algorithms.
The straight-line distance between two points is measured using a straightforward
distance metric. This metric is frequently employed when the distribution is some-
what uniform and the data are continuous.
– Rather than calculating the distance along a straight line, the Manhattan distance
metric calculates the total of the absolute differences between two places. It is
generally employed when the data is discrete and has a nonuniform distribution.
– The Minkowski distance metric, which combines the Manhattan and Euclidean
distance metrics, is another intricate distance metric. It is employed when the un-
derlying distribution can be roughly classified as either uniform or nonuniform
and the data is continuous.
– The maximum absolute difference between two points is measured using the Che-
byshev distance metric, a straightforward distance metric. It is frequently applied
to highly variable data that could contain anomalies or other unreliable data
pieces.
– A more intricate distance metric that accounts for the correlation between data
variables is the Mahalanobis distance metric. It is frequently applied to datasets
with numerous dependent variables correlated with one another.
– A straightforward distance metric known as the Hamming distance counts the
number of bits that differ between two binary integers. It is generally used when
there is no ordering for categorical data.
3 Classification and regression algorithms 71
– The cosine similarity distance metric, which is more frequently employed for text
processing applications, quantifies the angle between vectors. NLP activities like
sentiment analysis and document categorization are where it is most commonly
used.
– The ratio of the difference between two series of values is calculated using the
Bray-Curtis distance metric. It is typically used when evaluating ecological dis-
tance metrics to compare the composition of various organisms in a dataset.
K-NN algorithms provide various distance measures with unique properties and ap-
plications. Accurate outcomes with the K-NN method depend on selecting the appro-
priate distance metric for the task.
The best K-value must be chosen for an ML algorithm to be accurate and have a low
prediction error. K determines how many adjacent data points are used to produce a
prediction. Since more data points can be considered for each prediction with a
higher K-value, predictions are typically more accurate. The algorithm may become
too influenced by the constraints of the training data rather than the patterns of the
data as a whole, if a higher K-value is used. This is known as overfitting. The elbow
approach is a reliable way to find the ideal K-value. Finding the “elbow” point, or the
point at which the error rate starts to plateau, entails plotting a graph of the mean
error rate versus the K-value. This is an example of the excellent K-value, which offers
the optimum compromise between precision and overfitting. The dimensionality
problem should also be considered when choosing a K-value [24]. This is the phenom-
enon where more K-values (neighbors) are needed to represent each point, and more
qualities (variables) appropriately are present in the data set. Choosing a K-value that
considers the number of variables present in the data set is crucial because doing oth-
erwise can result in overfitting. Finally, domain expertise should also be considered
while selecting an outstanding K-value. For instance, a more considerable K-value
might be chosen in a medical dataset to produce a more precise forecast. The elbow
approach should be used to calculate the ideal K-value, considering the number of
variables in the data collection and the domain’s needs. These elements can be consid-
ered to select a K-value that strikes a balance between overfitting and precision.
Powerful ML algorithms like K-NN are frequently employed for supervised and unsu-
pervised tasks. K-NN is a straightforward yet effective algorithm that may be used for
various issues across numerous fields. In classification and regression problems, it is
72 G. Venkataramana Sagar et al.
Powerful ML algorithms like K-NN can identify outliers, solve regression and classifi-
cation problems, and recommend products. It can process various data kinds and pro-
duce reliable answers, primarily when the data is organized into complicated clusters
or contains many classes.
Due to its accuracy, the random forest has gained popularity as a method for vari-
ous applications. Ensemble approaches come in many different varieties. Bagging,
boosting, stacking, and blending are a few of these [25]. Bagging makes use of statistical
averaging and bootstrapping, which both serve to lessen overfitting. Boosting examines
the significance of incorrectly categorized data items and bolsters their predictions.
Stacking integrates many models, including neural networks and decision trees, to cre-
ate a more potent prediction model. Through blending, the predictions from each
model are combined and weighted according to how accurate they were. A robust ap-
proach for increasing the accuracy of models is ensemble methods. They are getting
increasingly popular because they provide a technique to build a more robust, special-
ized model out of a collection of weaker, varied models. They are used in various appli-
cations, such as medical ones, speech recognition, NLP, picture and text recognition,
and so on.
3.9.1 Bagging
The bagging algorithm creates ensembles of classifiers from a single base model. It
randomly selects features from a given dataset or sampling the training set. It then
uses the chosen features to train the base model multiple times, resulting in numer-
ous trained models. The predictions from each of these models are then combined to
give an overall better estimation or classification than any models can provide. This
algorithm can be used for any type of ML task, such as classification, regression, or
clustering, as long as it produces a set of labeled data. In particular, it is used for clas-
sification tasks because it can take multiple “weak” classifiers and combine them to
create a “strong” overall classification result. Bagging can reduce the variance of a
single classifier by combining the votes of multiple models. It is beneficial for complex
and nonlinear datasets. By using many different models trained on different subsets
of the training data, the variance of the predictions from each classifier is reduced.
Also, bagging can reduce the time required to train a model since it only requires
a smaller subset of the training data for each base model. Bagging is a powerful en-
semble learning method that can improve the accuracy of an ML model while retain-
ing low bias and managing its variance. It can make complex datasets more accurate,
provided that its base model is already best-suited for the problem. It is an essential
technique for improving the performance of an ML model and can be used for vari-
ous ML tasks such as classification, regression, and clustering.
3.9.2 Boosting
grating several low-accuracy models and boosting works by training each model inde-
pendently using various subsets of the same dataset. The algorithm introduces the mod-
els iteratively, with each new model concentrating on the flaws of the preceding
models. Each model is initially given equal weight, but the consequences are later
changed such that the most accurate models have a more significant impact on the en-
semble’s decisions. This indicates that the models improve and become more accurate
as the algorithm develops since each new model makes up for flaws in the preceding
one. Boosting can also be used for regression and extreme learning machines, although
it is frequently employed to increase accuracy in classification tasks like recognizing
spam emails. It often works with other ML methods, including decision trees, neural
networks, and SVMs. Boosting can be used to address a wide range of issues. It can be
used, for instance, to determine which features in a multidimensional dataset are the
most crucial [26]. It can also increase accuracy for applications like automatic text clas-
sification, object identification, and picture processing. The ensemble can create a ro-
bust system that can accurately classify or identify complicated patterns by integrating
several weak models. Combining soft models into strong ensembles using the ML ap-
proach of boosting is powerful and effective. It is an essential tool in the ML toolbox
and is frequently used to increase accuracy in challenging jobs.
3.9.3 Stacking
A powerful ML technique called the stacking ensemble method combines the output
from various models to provide a combined forecast that is more precise. With this
technique, data scientists can generate super-predictive models from weak learners
that can be mixed and stacked to produce a super-model that may be more accurate
than any of its component models. The stacking ensemble approach, in contrast to con-
ventional ML algorithms, uses “meta-learning” to improve the overall model based on
the outcomes of each fundamental model. It begins by training several basic models,
such as decision trees or simple regression, before calculating the weighted average of
all the forecasts to arrive at the final prediction. This approach enables data scientists
to identify minute details and variations in the data, resulting in greater accuracy. The
main benefit of the stacking ensemble method is that it allows for creating of highly
predictive models by combining the results of several learners. Since different learners
can be used to find various patterns or features within the data, it is also a valuable
tool for handling multiple sorts of data. Additionally, it combines many learning techni-
ques to outperform single models regarding predicted accuracy. The complexity of the
stacking ensemble approach is its main flaw. To train numerous learners, the data sci-
entist must devote time to model selection, feature engineering, hyperparameter opti-
mization, and other associated processes [27]. As a result, it takes longer and requires
more effort than traditional ML techniques. Combining numerous weak learners, the
stacking ensemble approach is a potent tool for creating accurate predictive models.
3 Classification and regression algorithms 75
The complexity of this method can make it difficult. Hence it is best employed by data
scientists with experience.
The results from numerous additional classifiers are combined using a voting classi-
fier and an ML algorithm to produce a consensus forecast. The voting classifier’s pre-
diction is made using the majority rule, where the class with the most votes is
selected as the output label. Voting classifiers have drawn interest in the ML commu-
nity because of their ease of use and potential to increase prediction accuracy over a
single classifier, theoretically. By merging several weak classifiers, the classifiers will
perform better on a given task than any single classifier. Voting classifiers can be ap-
plied to various studies, including classification, regression analysis, and anomaly de-
tection. Voting classifiers have the benefit of requiring little work from the user. The
majority of classifiers determine the forecast. Thus, the user can combine a variety of
classifiers to get a reasonably accurate prediction. Voting classifiers can quickly adapt
to new data. Therefore, they only require a little fine-tuning. Due to the variety of
classifiers used in the combination, voting classifiers tend to be more robust to noise
[28]. Better generalization performance is the consequence, which is advantageous
for data sets with high levels of unpredictability. Voting classifiers have certain down-
sides, nevertheless, in addition to these advantages. Their main drawback is that they
use more computer resources because multiple classifiers need to be trained and as-
sessed using various data sets. The majority vote will likewise be weak if the classifier
contains weak classifiers because they are only as good as their pieces. Additionally,
although voting classifiers might boost a weak classifier’s performance, feature-based
fusion is sometimes a preferable way to merge classifiers at a higher order level.
An ML technique called the ensemble technique enables the use of numerous models
in combination to produce more accurate predictions than the models used alone. En-
semble approaches aggregate the outcomes of various models for increased forecast
accuracy. Applications for ensemble methods include sentiment analysis, picture clas-
sification, image recognition, recommender systems, and many more. Sentiment anal-
ysis is one such application for ensemble methods. To increase the precision of
predictions in sentiment analysis, ensemble methods can combine the output of vari-
ous models. The production of numerous sentiment analysis models, such as those
based on NLP and ML algorithms, can be combined in a sentiment analysis system,
for example. This makes it easier to comprehend the sentiment expressed in an input
text document more precisely. Two more applications of ensemble methods are pic-
76 G. Venkataramana Sagar et al.
ture classification and image recognition [29]. Using image classification, models can
be combined to identify whether an image contains a specific object or scenario. Like
text recognition, image recognition algorithms can be integrated to identify and cate-
gorize items in photos accurately. Another practical application for ensemble algo-
rithms is recommender systems. Recommender systems use ensemble models to
aggregate the findings of several models and make user-specific product recommen-
dations. As a result, the algorithm may produce more precise proposals. Applications
for the ensemble methods include sentiment analysis, picture classification, image
recognition, and recommender systems, among many more. Ensemble methods can
produce extremely precise forecasts by combining the outcomes of various models.
Additionally, it can be used to find patterns, correlations, and trends in data that
may take time to be apparent. In various fields, from economics to biology, linear re-
gression is a flexible and necessary tool for data analysis [30]. It offers a handy
method of examining the connections between factors and may aid researchers in dis-
covering solutions to issues.
Regression analysis that uses polynomials to represent the connection between a sin-
gle dependent variable (y) and one or more independent variables (x) is known as
polynomial regression. This approach allows us to fit models representing the data as
linear, quadratic, cubic, or higher order polynomials. Polynomial regression might be
helpful when data display a nonlinear pattern, such as when an object is speeding up,
slowing down, or changing directions. It can also be used to examine situations when
it is challenging to choose the correct sequence for the polynomial to employ. In real-
ity, it can occasionally approximate a number that is not otherwise directly observ-
able, such as the rate of acceleration in issues involving motion. This method allows
us to fit a more accurate approximation of the connection than a linear regression
would.
Additionally, we can capture more specific information in our predictions using
higher order polynomials, which could lead to more precise estimations for the de-
pendent variable. Ensuring that our model is not overfitting the data is a crucial fac-
tor to consider while utilizing polynomial regression. When a model tries to account
for too much data variance, it is said to be overfitting; this can produce inaccurate
predictions and a poorer overall fit. Testing our model using cross-validation and
leveraging techniques to prevent this is crucial. A potent method, polynomial regres-
sion is beneficial when working with data that exhibits a nonlinear pattern. We can
capture more detail in our predictions and get a better fit by using higher order poly-
nomials. Overfitting can be problematic. Thus, it is advisable to use cross-validation
and leverage techniques to avoid it.
A supervised ML approach called support vector regression (SVR) is used for regres-
sion tasks. It is an SVM approach that transforms data into higher dimensional spaces
from which it may then use linear regression to identify nonlinear patterns. The pro-
gram divides the input data into two classes using hyperplanes and maps it to a
higher dimensional space. The support vectors are employed to choose the best hyper-
plane to execute linear regression and divide the data into classes. The SVR algorithm
can be applied to issues like projecting the price of a property or automobile, a stock
78 G. Venkataramana Sagar et al.
market trend, sales figures, etc. When data is not linear and linear regression cannot
be used to solve the problem, this method is appropriate. SVR is a versatile model that
works well to increase prediction accuracy.
Additionally, it can be used to implement regularization and lower variance for
more accurate outcomes quickly. SVR is less well-known than other ML methods, but
it provides a better data representation that can aid in making more accurate predic-
tions [31]. SVR is one of the most effective strategies for regression issues. In many
situations, it can perform better than conventional ML methods and contribute to gen-
erating more accurate outcomes. Due to this, it is essential to anticipate stock prices
and market trends in the financial industry.
A supervised learning method called decision tree regression can be used to forecast
continuous values. It operates by segmenting the input space into several parts (or
segments) and predicting the average data response in each portion. The decision pa-
rameters for the specific problem determine how the partitions are divided and what
values are assigned to the regions. Decision tree regression forecasts actual values
rather than classes like classification tree functions. The concept behind learning a
decision tree regression model to predict numerical values is to build a tree where
each node represents a decision point or a rule. The model traces the tree from root
to leaf to forecast the outcome. Decision tree regression is appropriate for a wide
range of issues since it can handle nonlinear data and accept category and numerical
characteristics.
A predictive model is fitted to a dataset using the random forest regression technique
to evaluate the associations between variables and make predictions. An ensemble
learning technique called a random forest combines numerous decision trees into a
single model, where each tree is a highly predictive model customized to a different
subset of the data. As it builds a model from labeled data (such as training data), the
random forest illustrates a supervised learning technique. A new data point can then
be added to the model without being labeled to forecast its class or value. Data science
activities like classification, clustering, and regression may all be performed using
random forest regression [32]. It can spot crucial details and connections between var-
iables, spot outliers and missing data, and forecast new data points.
3 Classification and regression algorithms 79
Continuous target variables are predicted using regression techniques from a given col-
lection of characteristics. Logistic and linear regression are the two most widely used
regression techniques. A constant target variable can be predicted using linear regres-
sion. It is employed to forecast an outcome in response to a predictor variable and to
clarify the connection between two or more variables. Logistic regression is used to esti-
mate the chance of an event occurring based on the values of a collection of predictor
variables. Finding the class to which a specific observation belongs is the objective of
this kind of classification problem. Other regression algorithms, such as polynomial,
neural networks, and SVR, exist in addition to linear and logistic regression. These tech-
niques can all be used to resolve many kinds of prediction issues [33]. Regression algo-
rithms have numerous typical applications and use cases, such as predicting customer
churn rates, server load, stock prices, marketing campaign performance, loan payback,
and fraud detection. These algorithms have been widely applied in engineering, mar-
keting, healthcare, and finance. Additionally, regression algorithms can be used to an-
ticipate future sales, consumer behavior trends, and product demand.
3.11.1 Cross-validation
Cross-validation sets aside a portion of the dataset while the model is trained using
the remaining data. This technique is used to assess predictive models. The model is
then evaluated using the held-out data. Every fold of the dataset is utilized as the test
set once throughout this process, which is performed numerous times. Therefore,
cross-validation enables you to gauge a model’s performance using hypothetical data.
In actual use, the estimate of model performance improves as the number of cross-
validation folds increases, but often 5–10 folds are utilized. Cross-validation has sev-
eral benefits, but its main advantages are that it is more accurate than a straightfor-
ward train-test split and more adaptable because it can be applied to a range of
sample sizes. It is also a popular way of fine-tuning model parameters, such as choos-
ing the ideal number of trees for a random forest.
Metrics for model evaluation are methods for measuring how well an ML model per-
forms against a collection of test or training data. These metrics are generally employed
to evaluate many models and choose the one that best satisfies a user’s needs. Accuracy,
precision, recall, F1 score, ROC (receiver operating characteristic)/AUC (area under the
curve), and log loss are standard model evaluation metrics. They aid in determining the
model’s suitability for a specific task. Different metrics should be employed depending
on the type of ML work; for instance, accuracy is appropriate for classification tasks,
whereas F1 score and log loss are more appropriate for ranking tasks.
3 Classification and regression algorithms 81
When a function is too closely fitted to a small number of data points, overfitting, a
modeling error, occurs. As a result, the model can correctly predict outcomes using
the training data set, but it cannot generalize and predict effects using new data.
Underfitting occurs when a statistical model cannot fully represent the underly-
ing structure of the data. It indicates that the model needs to fit the data adequately
and sufficiently account for much of the variation. Due to the model’s inability to ac-
count for the complexity of the data, underfitting causes erroneous predictions and a
significant bias [35].
Model selection strategies are collections of methods for selecting the best model from
a pool of candidate models. In addition to statistical criteria like cross-validation and
the Akaike information criterion, the procedures can include heuristics like the Bayes-
ian information criterion and stepwise regression and visual tools like parallel coordi-
nates and scree plots. In the end, bias and variance are traded off when choosing a
model. A model is said to be overfitting if its tendency is low but its conflict is high,
whereas it may be underfitting if its bias is high but its variance is low. The choice of
a model is a crucial stage in predictive modeling because it can significantly increase
its accuracy.
ML tools, such as classification and regression algorithms, are used to examine data
and create predictions. Classification algorithms are used to categorize data into dis-
tinct groups, such as assessing whether or not an email is spam. Regression algo-
rithms are employed to predict continuous quantities, such as the stock price or the
volume of website visits. Regression and classification algorithms are frequently com-
bined to create AI systems that are better capable of making precise predictions. Typi-
cal classification techniques include decision trees, logistic regression, and K-NN.
SVMs and polynomial regression are frequently used regression techniques [36].
3.12.3.1 Challenges
Overfitting, selection bias, data noise, and interpretability are a few of the problems
that classification and regression algorithms must deal with. An ML algorithm overfits
the training set when it fits the data too closely, producing models that do not general-
ize well to new data [38]. Selection bias might happen when the data used to build the
model do not represent the actual population. When there is a sufficient number of
valid training data or an excessive number of unimportant characteristics, data noise
may improve the model’s accuracy and interpretability.
3 Classification and regression algorithms 83
References
[1] Strecht, P., Cruz, L., Soares, C., Mendes-Moreira, J., & Abreu, R. (2015). A comparative study of
classification and regression algorithms for modelling students’ academic performance. Portugal:
International Educational Data Mining Society.
[2] Loh, W. Y. (2011). Classification and regression trees. Wiley Interdisciplinary Reviews: Data Mining
and Knowledge Discovery, 1(1), 14–23.
[3] Hoffmann, F., Bertram, T., Mikut, R., Reischl, M., & Nelles, O. (2019). Benchmarking in classification
and regression. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 9(5), e1318.
[4] Loh, W. Y. (2014). Fifty years of classification and regression trees. International Statistical Review,
82(3), 329–348.
[5] Pathak, S., Mishra, I., & Swetapadma, A. (2018, November). An assessment of decision tree based
classification and regression algorithms. In 2018 3rd International Conference on Inventive
Computation Technologies (ICICT) (pp. 92–95). IEEE.
[6] Zhou, Z. H., & Li, M. (2007). Semisupervised regression with cotraining-style algorithms. IEEE
Transactions on Knowledge and Data Engineering, 19(11), 1479–1493.
[7] Liaw, A., & Wiener, M. (2002). Classification and regression by random forest. R News, 2(3), 18–22.
[8] Loh, W. Y. (2008). Classification and regression tree methods. Encyclopedia of Statistics in Quality
and Reliability, 1, 315–323.
[9] Ponnusamy, M., Bedi, P., Suresh, T., Alagarsamy, A., Manikandan, R., & Yuvaraj, N. (2022). Design
and analysis of text document clustering using salp swarm algorithm. The Journal of
Supercomputing, 78(14), 16197–16213.
[10] Rocha, M., Cortez, P., & Neves, J. (2007). Evolution of neural networks for classification and
regression. Neurocomputing, 70(16–18), 2809–2816.
[11] Ren, Y., Zhang, L., & Suganthan, P. N. (2016). Ensemble classification and regression-recent
developments, applications and future directions. IEEE Computational Intelligence Magazine, 11(1),
41–53.
[12] Jaiswal, J. K., & Samikannu, R. (2017, February). Application of random forest algorithm on feature
subset selection and classification and regression. In 2017 World Congress on Computing and
Communication Technologies (WCCCT) (pp. 65–68). IEEE.
[13] Li, H., Sun, J., & Wu, J. (2010). Predicting business failure using classification and regression tree: An
empirical comparison with popular classical statistical methods and top classification mining
methods. Expert Systems with Applications, 37(8), 5895–5904.
[14] Mohammed, V. A., Mohammed, M. A., Mohammed, M. A., Logeshwaran, J., & Jiwani, N. (2023,
February). Machine learning-based evaluation of heart rate variability response in children with
84 G. Venkataramana Sagar et al.
autism spectrum disorder. In 2023 Third International Conference on Artificial Intelligence and
Smart Energy (ICAIS) (pp. 1022–1028). IEEE.
[15] Orabona, F., Crammer, K., & Cesa-Bianchi, N. (2015). A generalized online mirror descent with
applications to classification and regression. Machine Learning, 99, 411–435.
[16] Chou, P. A. (1991). Optimal partitioning for classification and regression trees. IEEE Transactions on
Pattern Analysis and Machine Intelligence, 13(04), 340–354.
[17] Deziel, M., Olawo, D., Truchon, L., & Golab, L. (2013, July). Analyzing the mental health of
engineering students using classification and regression. In Educational Data Mining 2013.
[18] Timofeev, R. (2004). Classification and regression trees (CART) theory and applications. Berlin:
Humboldt University, 54.
[19] Corrales, D. C., Ledezma, A., & Corrales, J. C. (2020). A case-based reasoning system for
recommendation of data cleaning algorithms in classification and regression tasks. Applied Soft
Computing, 90, 106180.
[20] Thomas, P., & Suhner, M. C. (2015). A new multilayer perceptron pruning algorithm for classification
and regression applications. Neural Processing Letters, 42, 437–458.
[21] Yang, Y. H., Lin, Y. C., Su, Y. F., & Chen, H. H. (2007, July). Music emotion classification: A regression
approach. In 2007 IEEE International Conference on Multimedia and Expo (pp. 208–211). IEEE.
[22] Rani, P., Verma, S., Yadav, S. P., Rai, B. K., Naruka, M. S., & Kumar, D. (2022). Simulation of the
Lightweight Blockchain Technique Based on Privacy and Security for Healthcare Data for the Cloud
System. International Journal of E-Health and Medical Communications (IJEHMC), 13(4), 1–15.
[23] Kannan, S., Yuvaraj, N., Idrees, B. A., Arulprakash, P., Ranganathan, V., Udayakumar, E., & Dhinakar,
P. (2021). Analysis of convolutional recurrent neural network classifier for COVID-19 symptoms over
computerised tomography images. International Journal of Computer Applications in Technology,
66(3–4), 427–432.
[24] Reddy, C. K., & Park, J. H. (2011). Multi-resolution boosting for classification and regression
problems. Knowledge and Information Systems, 29, 435–456.
[25] Khan, R. A., Ahmad, N., & Minallah, N. (2013). Classification and regression analysis of the prognostic
breast cancer using generation optimizing algorithms. International Journal of Computer
Applications, 68(25), 42–47.
[26] Ghasemain, B., Asl, D. T., Pham, B. T., Avand, M., Nguyen, H. D., & Janizadeh, S. J. V. J. O. E. S. (2020).
Shallow landslide susceptibility mapping: A comparison between classification and regression tree
and reduced error pruning tree algorithms. Vietnam Journal of Earth Sciences, 42(3), 208–227.
[27] Balda, E. R., Behboodi, A., & Mathar, R. (2019). Perturbation analysis of learning algorithms:
Generation of adversarial examples from classification to regression. IEEE Transactions on Signal
Processing, 67(23), 6078–6091.
[28] Liu, Q., Wang, X., Huang, X., & Yin, X. (2020). Prediction model of rock mass class using classification
and regression tree integrated AdaBoost algorithm based on TBM driving data. Tunnelling &
Underground Space Technology, 106, 103595.
[29] Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model
for efficient resource allocation in 5G broadband communication networks. ICTACT Journal On Soft
Computing, 13(3), 2986–2994.
[30] Yadav, S. P., & Yadav, S. (2019). Mathematical implementation of fusion of medical images in
continuous wavelet domain. Journal of Advanced Research in Dynamical and Control System, 10(10),
45–54.
[31] Priyam, A., Abhijeeta, G. R., Rathee, A., & Srivastava, S. (2013). Comparative analysis of decision tree
classification algorithms. International Journal of Current Engineering and Technology, 3(2),
334–337.
3 Classification and regression algorithms 85
[32] Zhang, L., Huettmann, F., Zhang, X., Liu, S., Sun, P., Yu, Z., & Mi, C. (2019). The use of classification
and regression algorithms using the random forests method with presence-only data to model
species’ distribution. MethodsX, 6, 2281–2292.
[33] Gunjan, V. K., Vijayalata, Y., Valli, S., Kumar, S., Mohamed, M. O., & Saravanan, V. (2022). Machine
Learning and Cloud-Based Knowledge Graphs to Recognize Suicidal Mental Tendencies.
Computational Intelligence and Neuroscience, 2022, 1–12
[34] Lou, Y., Caruana, R., & Gehrke, J. (2012, August). Intelligible models for classification and regression.
In Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and
Data Mining (pp. 150–158).
[35] Yang, T., Gao, X., Sorooshian, S., & Li, X. (2016). Simulating California reservoir operation using the
classification and regression‐tree algorithm combined with a shuffled cross‐validation scheme.
Water Resources Research, 52(3), 1626–1651.
[36] Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J. . . .
Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey
wolf optimization in wireless sensor networks. Security and Communication Networks, 2022, 1–13
[37] Grubinger, T., Zeileis, A., & Pfeiffer, K. P. (2014). evtree: Evolutionary learning of globally optimal
classification and regression trees in R. Journal of Statistical Software, 61, 1–29.
[38] Yusoff, M. H., Chrol-Cannon, J., & Jin, Y. (2016). Modeling neural plasticity in echo state networks for
classification and regression. Information Sciences, 364, 184–196.
Ashwin M., Nazeer Shaik, Faiz Akram, Getachew Mamo Wegari
4 Clustering and association algorithm
Abstract: This chapter provides an extensive examination of clustering and association
algorithms, offering a comprehensive understanding of these essential unsupervised learn-
ing techniques in the field of machine learning. The chapter commences with an introduc-
tory section that underscores the significance of unsupervised learning, and presents a
detailed overview of clustering and association algorithms. The various clustering algo-
rithms covered in the following sections are k-means, hierarchical clustering, density-
based clustering (like DBSCAN), and Gaussian mixture models. Additionally, the chapter
explores evaluation metrics for clustering, including both internal and external evaluation
metrics, and offers advise on how to properly interpret the evaluation results. The chapter
also examines dimensionality reduction methods, particularly principal component analy-
sis and t-SNE, and clarifies how to use them in clustering scenarios. The Apriori algorithm,
frequent itemset mining, association rule development, and methods for judging rule in-
terestingness are all covered in the section on association rule learning. Along with ex-
plaining their individual uses, the chapter also goes into advanced clustering methods like
density-based spatial clustering, mean-shift clustering, spectral clustering, and affinity
propagation. The chapter also discusses association rule mining, ensemble clustering, and
time series clustering in huge datasets. The chapter concludes with an overview of the
core ideas and developments in unsupervised learning, along with a list of the field’s pres-
ent research roadblocks and potential future paths.
4.1 Introduction
4.1.1 Overview of clustering and association algorithms
Data mining and ML employ the potent clustering and association algorithms. In
order to find underlying structures or patterns in the data, clustering algorithms put
comparable data points together according to their properties [1]. K-means, DBSCAN,
and hierarchical clustering are popular techniques for clustering data. These algo-
rithms analyze the proximity or similarity between data points to form clusters, en-
abling insights into the natural groupings present in the data, as in Figure 4.1.
On the other hand, association algorithms focus on identifying relationships or
associations between different items or variables in a dataset. Apriori is the most
well-known association algorithm; it finds frequent item sets and provides association
rules, based on support and confidence metrics. The co-occurrence of items or varia-
bles in the dataset is indicated by these rules.
https://doi.org/10.1515/9783111323749-004
88 Ashwin M. et al.
Large Data
set
Both clustering and association algorithms are crucial in exploratory data analysis,
customer segmentation, market basket analysis, and recommendation systems [2].
They provide valuable insights and enable decision-making based on patterns, rela-
tionships, and groupings present in the data.
Unsupervised learning [3] plays a crucial role in machine learning and has significant
importance for several reasons:
1. Discovering hidden patterns: Unsupervised learning algorithms enable the identi-
fication of hidden patterns and structures within data without the need for la-
beled examples. This can lead to valuable insights and discoveries, especially in
cases where the underlying patterns are unknown or difficult to define.
2. Data exploration and preprocessing: Unsupervised learning techniques provide
tools for data exploration, visualization, and preprocessing. Clustering algorithms,
for example, help identify natural groupings in the data, enabling a better under-
standing of the data’s characteristics and potential relationships.
3. Anomaly detection: Unsupervised learning is useful in detecting anomalies or out-
liers in datasets. By learning the normal patterns within the data, unsupervised
4 Clustering and association algorithm 89
algorithms can flag instances that deviate significantly from the expected behav-
ior, helping in fraud detection, fault diagnosis, and cybersecurity.
4. Dimensionality reduction: Numerous unsupervised learning algorithms, like prin-
cipal component analysis (PCA) and t-SNE, help to decrease the dimensionality of
high-dimensional data while maintaining its fundamental structure. This makes
subsequent supervised learning tasks more efficient and makes feature selection
and data visualization easier.
5. Recommendation systems: Unsupervised learning plays a vital role in building
recommendation systems. Collaborative filtering techniques, such as matrix fac-
torization and clustering-based methods, can analyze user behavior and preferen-
ces to provide personalized recommendations without requiring explicit labels or
ratings.
6. Data generation and synthesis: The distribution of the original data can be used
to create synthetic data using unsupervised learning methods. This is especially
beneficial when there is a lack of labeled data or when enhancing training data-
sets for supervised learning tasks.
A popular approach for dividing a dataset into K different clusters is K-means cluster-
ing. The final centroids represent the cluster centers, and each data point is associated
with the cluster that has the closest centroid. K-means is efficient and scalable for
large datasets, but it has limitations. It assumes a spherical shape of clusters, is sensi-
90 Ashwin M. et al.
tive to the initial centroid positions, and may converge to local optima. However, it
remains a popular algorithm for clustering tasks in various domains due to its sim-
plicity and effectiveness [5].
Discovering clusters based on the density of data points in a dataset is the goal of den-
sity-based clustering algorithms like DBSCAN (Density-Based Spatial Clustering of Ap-
plications with Noise). DBSCAN, in contrast to other clustering methods, does not
presume that clusters have a particular shape and can find clusters of any shape [7].
DBSCAN is capable of handling noise in the data and can identify clusters of vary-
ing density. It automatically determines the number of clusters based on the data char-
acteristics. Nevertheless, the performance of DBSCAN can be influenced by the choice
of its parameters, including ’“epsilon” and “minPts.”. Overall, DBSCAN is a valuable al-
gorithm for discovering clusters in complex and irregularly shaped datasets.
Probabilistic models, known as Gaussian mixture models (GMMs), are used to repre-
sent and cluster data that is believed to have been created from a combination of
Gaussian distributions. GMMs are widely used in various ML applications, including
clustering, density estimation, and generative modeling.
GMMs are flexible and can model complex distributions by combining multiple
Gaussian components. They can handle overlapping or elongated clusters in the data.
GMMs are useful for clustering tasks where the underlying data distribution is not
easily separable or when there is uncertainty in cluster assignment. However, GMMs
require the assumption that the data points are generated from Gaussian distribu-
tions, which may not hold in all cases.
4 Clustering and association algorithm 91
These are only a few of the numerous uses that clustering algorithms have. Clustering
is a useful technique in many disciplines because it can reveal patterns, groupings,
and linkages within data, assisting with decision-making, pattern detection, and data
exploration.
92 Ashwin M. et al.
Internal Evaluation
Metrics
Davies-Bouldin Calinski-Harabasz
Silhouette Score Dunn Index
Index Index
1. Silhouette score: The compactness and separation of clusters are assessed using
the Silhouette score. Better-defined clusters are indicated by a higher Silhouette
score.
2. Dunn index: The Dunn index evaluates the compactness and separation of
clusters.
3. Davies-Bouldin index: Based on the ratio of the average dissimilarity across clus-
ters to the greatest intra-cluster dissimilarity, the Davies-Bouldin Index evaluates
the clustering quality. Lower numbers represent more well-defined clusters.
4. Calinski-Harabasz index: To increase inter-cluster separation and reduce intra-
cluster variation, this index evaluates the ratio of between-cluster dispersion to
within-cluster dispersion. Higher Calinski-Harabasz scores indicate better cluster-
ing quality.
These internal evaluation metrics provide insights into the cohesion and separation
of clusters, helping to determine the effectiveness of clustering algorithms without re-
quiring external reference information. It is important to consider multiple metrics
and compare results across different algorithms or parameter settings for a compre-
hensive evaluation.
4 Clustering and association algorithm 93
External Evaluation
Metrics
Adjusted Rand
Rand Index (RI) F-measure Jaccard Coefficient
Index (ARI)
1. Rand index (RI): The RI measures the similarity between two clusterings by count-
ing the number of agreements (both in the same cluster or in different clusters)
and disagreements (one in the same cluster, the other in a different cluster) be-
tween the two clusterings. The index ranges from 0 to 1, where 1 indicates a per-
fect match.
2. Adjusted RI (ARI): The ARI accounts for the expected agreement between random
clusterings, and modifies the RI for chance agreement.
3. F-measure: To rate the accuracy of clustering results, the F-measure combines
precision and recall. By taking into account both the completeness of each cluster
and the purity of each class inside a cluster, it determines the harmonic mean of
the two measures.
4. Jaccard coefficient: Based on the intersection and union of data point assign-
ments, the Jaccard coefficient gauges the similarity between two clusterings.
hand side) and a consequent (the items on the right-hand side). These rules provide in-
sights into the dependencies and associations between items, enabling applications
such as market basket analysis, recommendation systems, and understanding customer
behavior in retail, e-commerce, and various domains [15].
A common approach used for frequent itemset mining in association rule learning is
the Apriori algorithm. It effectively finds common itemsets by using a breadth-first
search approach.
The process of identifying groups of items that frequently appear together in a dataset
is known as frequent itemset mining. The support calculation for itemsets is the pri-
mary equation utilized in frequent itemset mining.
To perform frequent itemset mining, the process typically involves the following steps:
Step 2. Pruning
Pruning is done by checking if all subsets of a candidate itemset are frequent.
Step 3. Iteration
The process iterates by generating larger candidate itemsets, counting their support,
and pruning until no more frequent itemsets can be found.
The frequent itemsets discovered through this process provide insights into the co-
occurrence relationships between items, which can be used to generate association
rules and understand patterns in the data.
Association rule generation involves deriving rules that express relationships be-
tween items based on the frequent itemsets discovered through frequent itemset min-
96 Ashwin M. et al.
ing. The primary equation used in association rule generation is the confidence calcu-
lation. Here is the equation for confidence:
Step 4. Pruning
Association rules are trimmed if they do not match the minimum confidence criterion
since they are viewed as being less significant or reliable.
The generated association rules provide insights into the dependencies and co-occurrence
relationships between items. They help in understanding the behavior and preferences of
customers, market basket analysis, cross-selling, and various decision-making processes.
In summary, association rule generation involves calculating the confidence of
association rules using the support of itemsets. By setting a minimum confidence
threshold, meaningful and significant rules can be extracted from the frequent item-
sets, providing valuable insights into item associations and patterns in the data.
Numerous disciplines can benefit from association rule mining, which has a wide
range of applications. Here are some typical use cases and applications for association
rule mining.
4 Clustering and association algorithm 97
Association rule mining is extensively used in retail and e-commerce for market bas-
ket analysis. It helps retailers identify frequently co-occurring items in customer
transactions, enabling them to understand purchasing patterns, recommend related
products, optimize store layouts, and implement effective cross-selling strategies.
Association rule mining is applied in web usage mining to understand patterns and
associations in user browsing behavior. It helps website owners optimize website lay-
outs, recommend relevant content, and personalize user experiences based on the
navigation patterns and preferences of website visitors.
Association rule mining is used in healthcare to analyze patient records, medical data-
sets, and clinical studies. It helps identify relationships between symptoms, diagnoses,
treatments, and patient outcomes. These insights assist in disease diagnosis, treatment
planning, drug discovery, and healthcare resource optimization.
When assessing the utility and impact of association rules in association rule mining,
several measures are commonly used [16, 17]. Here are three widely utilized measures:
Support:
Support quantifies how frequently or frequently an itemset appears in the dataset.
Support is calculated as the percentage of transactions in the dataset that contain
both A and B for an association rule A → B. It is calculated as given below and aids in
determining how frequently the rule occurs:
These additional measures provide different perspectives on the value of the associa-
tion rules, capturing various aspects such as implication strength, deviation from in-
dependence, similarity, and asymmetry between the antecedent and consequent.
Depending on the specific requirements and objectives of the analysis, these measures
can be utilized to evaluate and select the most interesting and relevant association
rules.
Graph theory and a similarity matrix’s eigenvectors are used in spectral clustering to
produce grouping. It clusters data points based on their connection patterns and
treats them as nodes in a graph. Nonlinearly separable data can be handled through
spectral clustering, which is especially useful for handling data with complicated
structures.
100 Ashwin M. et al.
Affinity propagation determines clusters by passing messages between data points to find
exemplars that best represent the clusters. It can discover the number of clusters automati-
cally and is suitable for applications like gene expression analysis and document clustering.
GMM assumes that the data points are generated from a mixture of Gaussian distribu-
tions. It models clusters as Gaussian distributions with different means and covarian-
ces. GMM can estimate the likelihood that a data point will belong to each cluster and
is effective for data with overlapping clusters.
Distance measures: Time series clustering often relies on distance measures that cap-
ture the similarity or dissimilarity between two time series. Examples include Euclidean
distance, dynamic time warping (DTW), and correlation-based measures.
Financial analysis: Clustering stock price or financial market data to identify groups
of similar time series for portfolio management or risk analysis
Sensor networks: Clustering sensor data to identify common patterns or behaviors for
environmental monitoring, smart grids, or Internet of Things (IoT) applications
Time series clustering helps in discovering meaningful patterns and structures within
time-dependent data, enabling better understanding, prediction, and decision-making
in various domains.
Ensemble clustering combines the basis clusterings into a single consensus clustering by
using integration methods. Majority voting, weighted voting, clustering co-occurrence,
and clustering agreement measures are common integration strategies.
Cluster stability analysis: Ensemble clustering often involves assessing the stability of
the clustering solutions to ensure the reliability of the ensemble result. Stability meas-
ures such as cluster-wise stability, pairwise similarity, or stability indexes are used to
evaluate the stability of clusters across different ensemble members.
Ensemble pruning and selection: Ensemble clustering may involve pruning or select-
ing a subset of base clusterings based on their quality or diversity measures to im-
prove the final clustering result and reduce computational complexity.
The evaluation of ensemble clustering aims to assess the quality and effectiveness of
the clustering results obtained from ensemble techniques. Here are some common
evaluation measures used for ensemble clustering:
– Jaccard coefficient: Measures the similarity between the clustering result and a
reference clustering solution
– Fowlkes-Mallows index: Calculates the geometric mean of the pairwise precision
and recall between the clustering result and a reference clustering
– Rand index: Compares the similarity of the clustering result with a reference clus-
tering by considering true positive and true negative pairwise agreements
– F-measure: Combines precision and recall to evaluate the clustering result against
a reference clustering
– Silhouette score: Based on the average intra-cluster distance and inter-cluster dis-
tance, evaluates the separation and compactness of clusters
Image and object recognition: Ensemble clustering is employed in image and object
recognition tasks to combine multiple feature extraction and clustering algorithms
for improved accuracy and robustness.
Bioinformatics: Ensemble clustering is used in genomic data analysis to identify gene expres-
sion patterns, discover co-expressed gene modules, and classify gene expression profiles.
Text mining: Ensemble clustering helps in text document clustering to group similar
documents for information retrieval, topic modeling, and document organization.
Social network analysis: Ensemble clustering is applied to analyze social networks and iden-
tify community structures, detect overlapping communities, and analyze network evolution.
Distributed computing: The calculation can be split over several nodes or machines
using distributed computing frameworks like Apache Hadoop or Apache Spark. Due
to this, processing may be done in parallel and computing resources can be used
effectively.
Incremental and online algorithms: Instead of processing the entire dataset at once,
incremental and online algorithms process the data in small chunks or as new data
arrives. This approach enables real-time analysis and reduces memory requirements.
Preprocessing and data cleaning: Large datasets often contain noisy or irrelevant data.
Proper data preprocessing, including data cleaning, noise removal, and outlier detec-
tion, can improve the quality and efficiency of clustering and association algorithms.
Parallel and distributed processing: To transfer the computational load across numer-
ous nodes or machines, use parallel and distributed computing frameworks like Apache
Spark or Apache Flink. Large dataset processing can be accelerated greatly as a result.
Indexing and approximation: Employ indexing techniques like k-d trees or R-trees to
speed up nearest neighbor searches or distance calculations in clustering algorithms.
Approximation algorithms can also be used to trade off accuracy for faster computa-
tion in large-scale scenarios.
106 Ashwin M. et al.
Stream processing: If dealing with streaming data, employ stream processing techniques
that allow for real-time clustering or association analysis. Algorithms like CluStream or
DStream can handle continuous data streams and provide up-to-date insights.
Scalable data storage: Efficiently store and retrieve large datasets using scalable data
storage systems like Apache Hadoop Distributed File System (HDFS), NoSQL data-
bases, or columnar databases. This enables seamless access to data during the cluster-
ing and association process.
Applying these techniques and considering the specific characteristics of large data-
sets can help address the computational, scalability, and efficiency challenges associ-
ated with clustering and association algorithms, enabling their effective application in
real-world scenarios.
Scalability: Developing algorithms and techniques that can efficiently handle ever in-
creasing volumes of data, including big data and streaming data
Advanced evaluation metrics: Designing novel evaluation metrics that capture the
quality, stability, and interpretability of clustering and association results, considering
specific application requirements
Integration with machine learning: Exploring the integration of clustering and associ-
ation techniques with other machine learning methods, such as deep learning or rein-
forcement learning, to enable more comprehensive data analysis
Privacy and security: Addressing privacy and security concerns in clustering and asso-
ciation mining, ensuring the protection of sensitive information and complying with
regulations
The future of clustering and association mining lies in advancing scalability, effi-
ciency, interpretability, and integration with other techniques. With ongoing research
and innovation, these methods will continue to provide valuable insights and support
decision-making across a wide range of applications and industries.
References
[1] Ahmad, A., & Khan, S. S. (2019). Survey of state-of-the-art mixed data clustering algorithms. IEEE
Access, 7, 31883–31902.
[2] Gurudath, S. (2020). Market basket analysis & recommendation system using association rules.
Master of Science in Big Data Management and Analytics, Griffith College, Dublin.
[3] Zhang, T., & Mao, S. (2020). Machine learning for end-to-end congestion control. IEEE
Communications Magazine, 58(6), 52–57.
108 Ashwin M. et al.
[4] Babenko, V., Panchyshyn, A., Zomchak, L., Nehrey, M., Artym-Drohomyretska, Z., & Lahotskyi,
T. (2021). Classical machine learning methods in economics research: Macro and micro level
example. WSEAS Transactions on Business and Economics, 18, 209–217.
[5] Sinaga, K. P., & Yang, M. S. (2020). Unsupervised K-means clustering algorithm. IEEE Access, 8,
80716–80727.
[6] Govender, P., & Sivakumar, V. (2020). Application of k-means and hierarchical clustering techniques
for analysis of air pollution: A review (1980–2019). Atmospheric Pollution Research, 11(1), 40–56.
[7] Chen, Y., Zhou, L., Bouguila, N., Wang, C., Chen, Y., & Du, J. (2021). BLOCK-DBSCAN: Fast clustering
for large scale data. Pattern Recognition, 109, 107624.
[8] Silvia Priscila, S., Sathish Kumar, C., Manikandan, R., Yuvaraj, N., & Ramkumar, M. (2022, February).
Interactive artificial neural network model for UX design. In International Conference on
Computing, Communication, Electrical and Biomedical Systems (pp. 277–284). Cham: Springer
International Publishing.
[9] Seem, A., Chauhan, A. K., Khan, R., & Yadav, S. P. (2020). Distributed artificial intelligence for
document retrieval. In: Satya Prakash Yadav, Dharmendra Prasad Mahato & Nguyen Thi Dieu Linh
(Eds.). Distributed Artificial Intelligence (pp. 59–68). Boca Raton: CRC Press.
[10] Zebari, R., Abdulazeez, A., Zeebaree, D., Zebari, D., & Saeed, J. (2020). A comprehensive review of
dimensionality reduction techniques for feature selection and feature extraction. Journal of Applied
Science and Technology Trends, 1(2), 56–70.
[11] Kiarashinejad, Y., Abdollahramezani, S., & Adibi, A. (2020). Deep learning approach based on
dimensionality reduction for designing electromagnetic nanostructures. npj Computational
Materials, 6(1), 12.
[12] Al-Turjman, F., Yadav, S. P., Kumar, M., Yadav, V., & Stephan, T. (Eds.). (2022). Transforming
management with AI, Big-Data, and IoT. India: Springer.
[13] Nithya, C., & Saravanan, V. (2018). A study of machine learning techniques in data mining.
International Scientific Refereed Research Journal, 1, 31–38.
[14] Praghash, K., Yuvaraj, N., Peter, G., Stonier, A. A., & Priya, R. D. (2023, May). Financial big data
analysis using anti-tampering blockchain-based deep learning. In Hybrid Intelligent Systems: 22nd
International Conference on Hybrid Intelligent Systems (HIS 2022), December 13–15, 2022
(pp. 1031–1040). Cham: Springer Nature Switzerland.
[15] Das, S., Kong, X., & Tsapakis, I. (2021). Hit and run crash analysis using association rules mining.
Journal of Transportation Safety and Security, 13(2), 123–142.
[16] Vougas, K., Sakellaropoulos, T., Kotsinas, A., Foukas, G. R. P., Ntargaras, A., Koinis, F., & Gorgoulis,
V. G. (2019). Machine learning and data mining frameworks for predicting drug response in cancer:
An overview and a novel in silico screening process based on association rule mining.
Pharmacology and Therapeutics, 203, 107395.
[17] Telikani, A., Gandomi, A. H., & Shahbahrami, A. (2020). A survey of evolutionary computation for
association rule mining. Information Sciences, 524, 318–352.
[18] Alashwal, H., El Halaby, M., Crouse, J. J., Abdalla, A., & Moustafa, A. A. (2019). The application of
unsupervised clustering methods to Alzheimer’s disease. Frontiers in Computational Neuroscience, 13, 31.
[19] Maharaj, E. A., D’Urso, P., & Caiado, J. (2019). Time series clustering and classification. New York:
CRC Press.
[20] Bañales, S., Dormido, R., & Duro, N. (2021). Smart meters time series clustering for demand
response applications in the context of high penetration of renewable energy resources. Energies,
14(12), 3458.
[21] Huang, D., Wang, C. D., Wu, J. S., Lai, J. H., & Kwoh, C. K. (2019). Ultra-scalable spectral clustering and
ensemble clustering. IEEE Transactions on Knowledge and Data Engineering, 32(6), 1212–1226.
[22] Wu, W. T., Li, Y. J., Feng, A. Z., Li, L., Huang, T., Xu, A. D., & Lyu, J. (2021). Data mining in clinical big data:
The frequently used databases, steps, and methodological models. Military Medical Research, 8, 1–12.
Nazeer Shaik, S. Chandra Sekaran, Amit Mahajan, Balkeshwar Singh
5 Reinforcement learning
Abstract: The field of reinforcement learning (RL) is introduced in this chapter, which
also looks at several RL techniques. The main goal of RL is to provide algorithms that
let agents discover the best policies through interactions with their surroundings
while maximizing cumulative rewards. In the first part of the chapter, Markov deci-
sion processes (MDPs), which provide a mathematical foundation for modeling RL
problems, are discussed. We look at value iteration and policy iteration as iterative
approaches to addressing MDPs. To help you find the ideal action-value function, we
present Q-Learning, an off-policy model-free RL algorithm. Deep Q-networks (DQNs),
which combine Q-learning with deep neural networks, are also addressed in order to
handle high-dimensional state spaces. Policy gradient methods are presented as an
alternative approach that directly optimizes policy parameters using gradient ascent.
Proximal policy optimization (PPO), a leading policy gradient algorithm, is discussed
for its ability to balance stability and policy performance. The chapter concludes by
emphasizing the significance of RL methods in training agents to make sequential de-
cisions in complex environments across various domains.
5.1 Introduction
The core of the subject of reinforcement learning (RL), a subsection of machine learn-
ing, is the investigation of agents that interact with their environment and discover
how-to-do behaviors that maximize a cumulative reward signal. This chapter will ex-
amine numerous RL techniques that let agents discover the best strategies through
making mistakes.
RL methods provide powerful tools for training agents to make sequential deci-
sions in complex environments. From classical methods like value iteration and policy
iteration to more recent advances like deep Q-networks and policy gradient algo-
rithms, RL continues to evolve and find applications in various domains, including
robotics, game playing, and autonomous systems [1, 2].
https://doi.org/10.1515/9783111323749-005
110 Nazeer Shaik, S. Chandra Sekaran, Amit Mahajan, Balkeshwar Singh
States (S):
States represent the configurations or situations an agent can encounter during the
decision-making process. The set S encompasses all possible states, and each state s ∈
S captures relevant information about the environment.
Actions (A):
Actions refer to the choices an agent can make in a given state. The set A represents
the available actions, and an action a ∈ A represents a specific decision the agent can
take.
Rewards (R):
The agent receives instant input from rewards, which influences how it learns. R(s, a, s’)
denotes the reward received when moving from state s to state s’ following the comple-
tion of action a.
Initialization:
Set a random beginning value for all states for the value function V(s) from s ∈ S.
5 Reinforcement learning 111
Iterative Update:
The value function can be updated by selecting the action that maximizes the action-
value function:
Convergence Check:
In order to attain the maximum number of iterations, keep repeating step 2 until the
change in the value estimations becomes minimal.
The value iteration algorithm utilizes the Bellman equation, which states that the
value of a state is equal to the maximum expected return achievable from that state:
Here, the ideal state-value function is represented by V✶(s), while the ideal action-
value function is represented by Q✶(s, a).
The algorithm modifies the value function after each iteration using the most recent
estimates of the action-value function. When calculating the action-value function, the tran-
sition probabilities T(s, a, s’) and immediate rewards R(s, a, s’) linked to the state-action
pairs are taken into consideration. The discount factor is used to rank future benefits [4].
Initialization:
Initialize the policy π arbitrarily for all states s ∈ S.
Policy evaluation:
Conduct a policy evaluation to calculate the value function Vπ(s) for the present policy
π. Continue iterating until the value estimations converge:
112 Nazeer Shaik, S. Chandra Sekaran, Amit Mahajan, Balkeshwar Singh
Policy improvement:
Based on the current value function estimates Vπ(s), update the policy π by selecting
the action that maximizes the expected return:
Update the policy by selecting the action that maximizes the action-value function:
Convergence check:
Until the policy changes or the allotted number of iterations is reached, repeat steps 2
and 3 as necessary.
During policy iteration, the value function V(s) is estimated by taking into account the
transition probabilities T(s, a, s’) and immediate rewards R(s, a, s’) connected to the
state-action pairings. The significance of future awards is determined by the discount
factor [6].
Policy iteration enables us to locate the best policy that maximizes the predicted
cumulative rewards in the MDP by repeatedly undertaking policy evaluation and pol-
icy improvement [7–9].
Model-based methods:
There are model-based approaches that develop models of the dynamics of the envi-
ronment, even though the majority of DRL algorithms are model-free. These models
can be used for planning, data augmentation, or improving sample efficiency in learn-
ing [12].
Rainbow:
Rainbow is a combination of various DRL techniques to achieve state-of-the-art per-
formance in value-based RL. It incorporates enhancements such as prioritized experi-
ence replay, distributional Q-learning, n-step returns, dueling networks, and NoisyNet
exploration to improve sample efficiency and stability.
DRL has demonstrated significant advancements in various domains and has the po-
tential to address complex real-world problems. However, training deep RL agents
often requires substantial computational resources, careful hyperparameter tuning,
and addressing challenges such as sample efficiency, exploration-exploitation trade-
offs, and generalization to new environments. Ongoing research aims to tackle these
challenges and further advance the capabilities of DRL algorithms.
Initialization:
Start by initializing the agent’s policy and value estimates based on prior knowledge
or random initialization.
Select action:
When faced with a state, the agent needs to select an action. This decision-making
process is guided by the exploration strategy [14].
a. Epsilon-greedy:
The epsilon-greedy approach is a straightforward and popular method. The majority
of the time, it includes choosing the action with the highest estimated value (exploita-
tion), but with a slim chance of doing so (epsilon) it entails choosing a random action
(exploration).
b. Boltzmann exploration:
Boltzmann exploration, also known as softmax exploration, selects actions probabilis-
tically based on their estimated values. A stochastic exploration technique is possible
since the likelihood of choosing an action is inversely correlated with its exponential
magnitude.
d. Thompson sampling:
A distribution over the value estimates for each action is maintained using the proba-
bility-based exploration technique known as sampling. By taking a value from the dis-
tribution for each action and sampling it, it selects the action with the greatest
sampled value.
e. Noise injection:
In some cases, random noise can be injected into the action selection process to en-
courage exploration. This can involve adding Gaussian noise to the action values or
applying random perturbations to the policy parameters.
Action selection:
Based on the chosen exploration strategy, select an action to take in the current state.
This action will determine the agent’s interaction with the environment.
5 Reinforcement learning 115
Environment interaction:
Execute the chosen action in the environment, then see the state transition and re-
ward that result.
For the agent to investigate and learn from the environment, exploration tactics are
essential. By intelligently balancing exploration and exploitation, these strategies en-
able the agent to discover optimal actions and states, leading to effective decision-
making and learning in RL tasks.
Action representation:
In continuous action spaces, the actions are typically represented as continuous varia-
bles. For simplicity, let us consider a single-dimensional continuous action space.
Policy representation:
Parameterized functions, such as neural networks, are frequently used to map states
to action values while learning a policy in continuous action spaces. The parameters
that define the probability distribution across the continuous action space are pro-
duced by the policy function π(s; θ) from the state s as an input [15].
Actor-critic methods:
The parameters of both networks are changed by employing methods like the deter-
ministic policy gradient or advantage-based techniques [16, 17].
116 Nazeer Shaik, S. Chandra Sekaran, Amit Mahajan, Balkeshwar Singh
Exploration strategies:
In continuous action spaces, exploration is particularly challenging due to the infinite
set of possible actions. Exploration strategies like adding noise to the policy parame-
ters (e.g., Ornstein-Uhlenbeck process) or using action space exploration techniques
(e.g., parameterizing the policy with a random variable) can be employed to encour-
age exploration and prevent the agent from being stuck in local optima.
These are some of the common approaches used to handle continuous action spaces
in RL. By using techniques such as deterministic policy gradients, function approxi-
mation, and actor-critic architectures, agents can effectively learn policies in continu-
ous action spaces.
Model learning:
In MBRL, the agent learns a model of the environment dynamics, which captures how
the states and rewards evolve over time.
Model-based planning:
Once the model is learned, the agent can use it for planning. Planning involves using
the model to simulate future trajectories and optimize the expected cumulative re-
ward. Action sequences that optimize the anticipated return can be created using
planning techniques like Monte Carlo tree search or model predictive control.
Model-based control:
Model-based control is the method used in MBRL to choose actions in the real world
based on the learnt model. The agent runs simulations of potential action sequences
using the model, analyzing the expected returns to make judgments.
5 Reinforcement learning 117
Model-based exploration:
In MBRL, the learned model can also be used for exploration. By simulating possible
actions and their outcomes using the model, the agent can explore regions of the
state-action space that it has not yet encountered. Various exploration strategies dis-
cussed earlier can be employed in conjunction with model-based methods to encour-
age exploration.
MBRL uses a variety of equations and algorithms, depending on the particular method
and problem area. MBRL has benefits including sample effectiveness and the capacity
for planning and environmental reasoning. However, it also requires accurate model
learning and planning algorithms that can handle the complexities of the environ-
ment dynamics. For effective and efficient RL, current research is focused on enhanc-
ing model learning methods, creating reliable planning algorithms, and combining
model-based and model-free approaches.
Cooperative MARL:
Cooperative MARL focuses on situations where agents aim to maximize a shared
global reward. The agents work together, sharing information and coordinating their
actions to achieve a common goal. Since one agent’s activities affect the outcomes and
rewards of other agents, learning effective coordination procedures can be challeng-
ing. The agents may employ centralized training with decentralized execution, where
a central learner aids in guiding the agents during training but they act independently
during execution.
Competitive MARL:
In competitive MARL, agents are pitted against each other, and their objectives are
adversarial or conflicting. Each agent seeks to maximize its own reward or perfor-
mance metric, often at the expense of others. Examples include competitive games
like chess or multiplayer video games. The agents need to learn strategies to outsmart
their opponents, anticipating their actions and finding optimal decision-making poli-
cies in the face of competition.
118 Nazeer Shaik, S. Chandra Sekaran, Amit Mahajan, Balkeshwar Singh
Communication-based MARL:
Communication-based MARL involves agents that can exchange information or mes-
sages with each other during their interactions. Communication enables agents to
share knowledge, coordinate actions, and cooperate more effectively. It allows for
higher levels of coordination and can help overcome challenges posed by complex en-
vironments. Agents can learn to communicate through explicit message passing or
through learned communication protocols.
Hierarchical MARL:
Hierarchical MARL involves learning policies at different levels of abstraction. Agents
learn both low-level policies that directly map states to actions and high-level policies
that determine goals or subtasks. High-level policies direct the agent’s decision-
making, and low-level policies manage the specifics of action selection. Hierarchical
approaches can simplify learning and decision-making in complex environments, en-
abling agents to tackle larger problems effectively.
Decentralized MARL:
Decentralized MARL uses local observations to allow each agent to independently
learn its own policy. The global state and comprehensive details of other agents’ poli-
cies are not available to agents. They make decisions based on their limited local ob-
servations, often resulting in partial observability. Decentralized MARL requires
agents to learn effective individual policies while considering the interactions with
other agents in the environment.
Centralized MARL:
In contrast to decentralized MARL, centralized MARL involves agents with access to a
centralized controller that provides global information about the environment and
other agents. The central controller can gather and process information from all
agents and provide guidance to optimize the overall performance. This type of MARL
can enhance coordination and decision-making, but it also raises challenges related to
communication overhead and scalability.
These types of MARL represent different problem settings and objectives in multi-
agent environments. Researchers employ a variety of algorithms and techniques to
address the complexities of interactions among agents, including value-based meth-
5 Reinforcement learning 119
ods, policy gradients, opponent modelling, game theory, and more. The choice of
MARL type and algorithm depends on the specific problem, the nature of interactions.
Robotics:
MARL is extensively used in robotics for multirobot coordination and collaboration. It
enables teams of robots to work together on tasks such as search and rescue, swarm
robotics, cooperative transportation, and multirobot exploration. MARL allows robots
to communicate, share information, and coordinate their actions to achieve common
objectives efficiently.
Autonomous vehicles:
MARL is valuable for developing intelligent systems for autonomous vehicles. Agents
representing vehicles can learn to navigate complex traffic scenarios, optimize traffic
flow, and handle coordination issues like merging, lane changing, and intersection
management. MARL can improve traffic efficiency, reduce congestion, and enhance
overall road safety.
Multiagent games:
MARL is extensively applied to competitive and cooperative games, both in video
games and real-world sports. It enables game agents to learn optimal strategies, adap-
tive behavior, and decision-making in dynamic and uncertain environments. MARL
has been used in games like poker, chess, Dota 2, and autonomous drone racing,
achieving impressive performance against human players.
Social networks:
MARL is employed in modeling and analyzing social networks to understand the dynam-
ics of social interactions. It can help in predicting user behavior and influence maximiza-
tion, recommendation systems, and online advertising. MARL enables agents to learn
and adapt to the evolving preferences and behaviors of individuals in the network.
Energy management:
MARL is beneficial for optimizing energy consumption and resource allocation in
smart grids and energy systems. Multiple agents, such as power generators, consum-
ers, and storage units, can learn to coordinate and balance energy supply and demand
efficiently. MARL can help reduce energy costs, improve renewable energy integra-
tion, and enhance the overall stability and reliability of the grid.
120 Nazeer Shaik, S. Chandra Sekaran, Amit Mahajan, Balkeshwar Singh
Multirobot simulations:
MARL is used to simulate multirobot systems and test different coordination and con-
trol strategies. Simulations help in evaluating system performance, scalability, and ro-
bustness before deploying the algorithms on real robots. MARL allows researchers to
investigate various scenarios, such as disaster response, environmental monitoring,
and industrial automation.
These are just a few examples of the broad range of applications where MARL techni-
ques are being explored and deployed. MARL offers the potential to address complex
problems that involve interactions among multiple intelligent agents, leading to im-
proved efficiency, coordination, and decision-making in various domains.
Scalability:
Traditional MARL algorithms have trouble handling large numbers of agents since
the state-action space is expanding exponentially and interactions are becoming more
complicated. The creation of scalable algorithms with effective learning and coordina-
tion capabilities in contexts with numerous agents is a future direction.
veloping methods for safe exploration, robustness to uncertain environments, and en-
suring the compliance of learned policies with safety constraints and regulations.
These challenges and future directions highlight the ongoing research efforts in MARL
to overcome limitations and improve the applicability and effectiveness of multiagent
systems. By addressing these challenges, MARL has the potential to revolutionize vari-
ous domains, enabling intelligent collaboration, decision-making, and coordination
among multiple agents in complex and dynamic environments.
References
[1] Foerster, J., Assael, Y. M., de Freitas, N., & Whiteson, S. (2016). Learning to communicate with deep
multi-agent reinforcement learning. Advances in Neural Information Processing Systems, 26,
2137–2145.
[2] Lowe, R., Wu, Y., Tamar, A., Harb, J., Abbeel, P., & Mordatch, I. (2017). Multi-agent actor-critic for
mixed cooperative-competitive environments. Advances in Neural Information Processing Systems,
30, 6379–6390.
[3] Sunehag, P., Lever, G., Gruslys, A., Vinyals, O., Spiessens, A., Mohanty, S., & Whiteson, S. (2018).
Value-decomposition networks for cooperative multi-agent learning. Proceedings of the AAAI
Conference on Artificial Intelligence, 32(1), 1–14.
[4] Rashid, T., Samvelyan, M., & Nowé, A. (2018). Qmix: Monotonic value function factorisation for deep
multi-agent reinforcement learning. In Thirty-Second AAAI Conference on Artificial Intelligence.
[5] Yu, H., Gupta, A., Xu, Z., Levine, S., & Abbeel, P. (2019). MOCA: Model-based offline contextual bandit
for continuous actions. In International Conference on Learning Representations.
[6] Omidshafiei, S., Pazis, J., Amato, C., How, J. P., & Vian, J. (2019). AlphaStar: Mastering the real-time
strategy game StarCraft II. Proceedings of the 38th International Conference on Machine Learning,
PMLR 139, 2021, China, pp.1–15
[7] Foerster, J. N., Farquhar, G., Afouras, T., Nardelli, N., & Whiteson, S. (2019). Counterfactual multi-
agent policy gradients. In International Conference on Machine Learning (pp. 2046–2054).
[8] Zheng, Y., Tang, H., Chen, H., Xu, Z., Hu, W., & Huang, Z. (2020). Coma: Communication efficient
multiagent reinforcement learning with actor-critic methods. Proceedings of the AAAI Conference
on Artificial Intelligence, 34(04), 3962–3969.
[9] Iqbal, S., & Sha, F. (2019). Actor-attention-critic for multi-agent reinforcement learning. In
International Conference on Machine Learning (pp. 2985–2994).
[10] Son, J., Kim, J., Choi, H., & Kim, S. (2019). Qtran: Learning to factorize with transformation for
cooperative multi-agent reinforcement learning. Proceedings of the AAAI Conference on Artificial
Intelligence, 33(01), 4282–4289.
[11] Rashid, T., Samvelyan, M., Van hasselt, H., Silver, D., & Nowé, A. (2020). Off-policy multi-agent dec-
POMDP planning with deep value iteration. In International Conference on Machine Learning
(pp. 7921–7931).
[12] Matsuo, Y., LeCun, Y., Sahani, M., Precup, D., Silver, D., Sugiyama, M., & Morimoto, J. (2022). Deep
learning, reinforcement learning, and world models. Neural Networks, 152, 267–275.
[13] Brunke, L., Greeff, M., Hall, A. W., Yuan, Z., Zhou, S., Panerati, J., & Schoellig, A. P. (2022). Safe
learning in robotics: From learning-based control to safe reinforcement learning. Annual Review of
Control, Robotics, and Autonomous Systems, 5, 411–444.
5 Reinforcement learning 123
[14] Ladosz, P., Weng, L., Kim, M., & Oh, H. (2022). Exploration in deep reinforcement learning: A survey.
Information Fusion, 85, 1–22.
[15] Fawzi, A., Balog, M., Huang, A., Hubert, T., Romera-Paredes, B., Barekatain, M., & Kohli, P. (2022).
Discovering faster matrix multiplication algorithms with reinforcement learning. Nature, 610(7930),
47–53.
[16] Degrave, J., Felici, F., Buchli, J., Neunert, M., Tracey, B., Carpanese, F., & Riedmiller, M. (2022).
Magnetic control of tokamak plasmas through deep reinforcement learning. Nature, 602(7897),
414–419.
[17] Wurman, P. R., Barrett, S., Kawamoto, K., MacGlashan, J., Subramanian, K., Walsh, T. J., & Kitano,
H. (2022). Outracing champion Gran Turismo drivers with deep reinforcement learning. Nature,
602(7896), 223–228.
[18] Wu, D., Lei, Y., He, M., Zhang, C., & Ji, L. (2022). Deep reinforcement learning-based path control and
optimization for unmanned ships. Wireless Communications and Mobile Computing, 2022, 1–8.
[19] Le, H., Wang, Y., Gotmare, A. D., Savarese, S., & Hoi, S. C. H. (2022). Coderl: Mastering code
generation through pretrained models and deep reinforcement learning. Advances in Neural
Information Processing Systems, 35, 21314–21328.
[20] Seno, T., & Imai, M. (2022). d3rlpy: An offline deep reinforcement learning library. The Journal of
Machine Learning Research, 23(1), 14205–14224.
[21] Yuvaraj, N., Karthikeyan, T., & Praghash, K. (2021). An improved task allocation scheme in serverless
computing using gray wolf Optimization (GWO) based reinforcement learning (RIL) approach.
Wireless Personal Communications, 117, 2403–2421.
[22] Silvia Priscila, S., Sathish Kumar, C., Manikandan, R., Yuvaraj, N., & Ramkumar, M. (2022, February).
Interactive artificial neural network model for UX design. In International Conference on
Computing, Communication, Electrical and Biomedical Systems (pp. 277–284). Cham: Springer
International Publishing.
[23] Awasthi, S., Naruka, M. S., Yadav, S. P., & De Albuquerque, V. H. C. (Eds.). (2023). AI and IoT-based
intelligent health care & sanitation. U.A.E: Bentham Science Publishers.
[24] Nayak, R. K., Tripathy, R., Saravanan, V., Das, P., & Anguraj, D. K. (2020, March). A novel strategy for
prediction of cellular cholesterol signature motif from G protein-coupled receptors based on rough
set and FCM algorithm. In 2020 Fourth International Conference on Computing Methodologies and
Communication (ICCMC) (pp. 285–289). IEEE.
[25] Sivapriya, T. R., & Saravanan, V. (2011). Automatic brain MRI mining using support vector machine
and decision tree. CiiT International Journal of Artificial Intelligent Systems and Machine Learning,
3(2), 109–116.
Syed Ibad Ali, Padmavathi Thasappan Kalaivaani, Ambigaipriya S.,
MD Rafeeq
6 Evaluation of AI model performance
Abstract: Artificial intelligence (AI) models are becoming more complex as a result of
recent advances in AI technology. As a result, there are numerous approaches to as-
sess the effectiveness of an AI model. In order to evaluate a model’s performance,
both its accuracy and how quickly it processes fresh data are measured. Examining
the model’s robustness to data that is not in the training set and its capacity for so-
phisticated problem solving are also included in the evaluation process. Because it
identifies opportunities for improvement and facilitates better decision-making, in-
terpretability is a crucial component of evaluating an AI model’s performance. Fi-
nally, a range of measures and methods are used in performance evaluation of AI
models to assess their efficacy.
6.1 Introduction
The effectiveness of artificial intelligence (AI) models can be evaluated using a variety
of metrics, such as accuracy, recall, precision, specificity, and F1-score. How accurately
a model can classify an input serves as a gauge of its accuracy. Recall gauges how well
a model can identify all pertinent instances of a particular class [1]. Precision evalu-
ates the model’s capacity to provide pertinent results for a specific query. The model’s
accuracy in identifying only pertinent results in response to a given query is mea-
sured by specificity. The F1-score, which combines recall and precision, serves as a
measure of the model’s overall performance. To create an AI model that performs
well, evaluation criteria should be selected based on the task at hand, and evaluated
frequently.
https://doi.org/10.1515/9783111323749-006
126 Syed Ibad Ali et al.
A table called a confusion matrix is frequently used to show how well a classification
model – also known as a “classifier” – performs on a set of test data for which the
true values are known. It makes it possible to see how an algorithm performs. The
confusion matrix is composed of four cells, each of which counts the number of true
positive (TP), true negative (TN), false positive (FP), and false negative (FN) values.
– TPs are the occurrences in which the model correctly predicted the positive class.
– TNs are those instances that occur when the model correctly predicts the negative
class.
– FPs are situations where the model anticipated the positive class wrongly.
– FNs are cases where the model predicted the negative class wrongly.
6 Evaluation of AI model performance 127
The confusion matrix provides for a more thorough examination than is possible
when only looking at the accuracy score, making it a useful tool for evaluating the
effectiveness of a classification algorithm. It enables the user to comprehend the mod-
el’s mistakes and how to enhance performance.
Two important metrics are utilized to assess a diagnostic test’s effectiveness: sensitiv-
ity and specificity. The percentage of true positive tests (those accurately recognized
as having the target disease) out of all individuals with the condition is referred to as
sensitivity [5]. The percentage of true negative tests (those accurately recognized as
not having the target condition) out of all those without the condition is what is re-
ferred to as specificity. Most people with the condition will be accurately identified by
a test with high sensitivity, and most people without the condition will be correctly
identified by a test with high specificity.
A table known as a confusion matrix is typically used to show how well a classifica-
tion model (also known as a “classifier”) performs on a set of test data for which the
true values are known. It makes it possible to see how well an algorithm performs.
The cases in each row of the matrix, as opposed to the examples, correspond to the
actual classes, whereas the columns of the matrix, or vice versa, belong to the antici-
pated classes [6]. The name confusion matrix was created because it makes it simple
to see when the system confuses two classes (i.e., when one is regularly referred to as
another).
The confusion matrix can be used to calculate a variety of metrics, such as preci-
sion, recall, accuracy, and the F1-score.
– Accuracy: This is determined by dividing the percentage of correct forecasts by
all other guesses.
– Precision: Divide the total number of true positives by the total number of TPs
and FPs to get at this.
– Bear in mind that this corresponds to the sum of the TPs and FNs.
– F1-score: A precision and recall harmonic mean is represented by this score. The
F1-score is considered to be the most helpful indicator when comparing models.
128 Syed Ibad Ali et al.
Interpreting the results of a predictive model and assessing its performance is a pro-
cess that involves understanding the concepts of accuracy, precision, and recall in
order to make an informed decision on whether a given predictive model is effective
or not [7]. Precision indicates how often the model can predict the outcome accu-
rately; recall is how many of the positive examples were properly identified; and ac-
curacy measures how closely the predictions match the actual values. By taking into
account all of these factors, one can evaluate the performance of a predictive model
and make an informed decision on whether to use it or not.
A measure of how well two parties agree on a subject is known as the Cohen’s kappa
coefficient, also known as the Cohen’s kappa statistic or Cohen’s kappa score. It is fre-
quently employed when attempting to determine the level of agreement between two
or more raters or observers on the same items in a collection. The kappa coefficient
takes into account the agreement that can be expected by random chance, thus help-
ing to eliminate any potential bias [9]. The result ranges from −1 (total disagreement)
to 1 (total agreement). A score of 0 indicates that the raters are in agreement only to
the extent expected by chance.
account both the number of agreements and disagreements between the raters while
also accounting for the possibility of agreement or disagreement by chance. Higher
values of the kappa statistic, which ranges from 0 to 1, imply greater agreement than
would be predicted by chance. Other methods, such as intra-class correlation coeffi-
cients, can also provide measures of inter-rater agreement.
Kappa values are a measure of agreement between two or more people who are rat-
ing the same item or event. They measure how well two judges agree with each other
on a particular rating, with higher values representing greater consistency. Numerous
situations, including employee evaluations, consumer satisfaction surveys, legal pro-
ceedings, and medical diagnoses, make use of kappa values [11]. The value is also
often used to measure inter-rater reliability.
130 Syed Ibad Ali et al.
6.4.1 Introduction to ROC curves and area under the curve (AUC)
Using ROC curves is a popular method for determining how well categorization mod-
els perform. The ROC curve is a graph that contrasts, on the x-axis, the true positive
rate for a specific model with the FPR. The ROC curve is used to gauge how well a
classifier can differentiate between real and fake positives (positive and negative clas-
ses). A typical statistic to evaluate a classifier’s prediction accuracy is the AUC.
To plot a ROC curve, the concept is simple. To start off, you must have a set of
data (or a model) that has been evaluated over a range of different thresholds (or de-
cision boundaries). Determine the ratio of true positives to FPs for each level. Then,
plot these places on a chart and draw a line that connects them. The graph that fol-
lows is the ROC curve [14].
Interpreting the results of a ROC curve is somewhat subjective, depending on the
context, but it is nonetheless useful for gauging the performance of a model. The ex-
pected outcome for a random classifier is that a model with an AUC score closer to 0.5
has better predictive power than a model with an AUC value closer to 1, on average.
Although two models may have comparable AUC values, this does not always imply
6 Evaluation of AI model performance 131
that they will perform similarly; other metrics like precision and recall can be utilized
to better understand the model.
An essential step in choosing the best classifier for a particular job is assessing a bi-
nary classifier’s performance. It entails judging how well the classifier can put data
points into one of two categories. A binary classifier’s performance can be assessed
using its accuracy, precision, recall, and F1-score.
– Accuracy is the proportion of data points among all the data points that were cor-
rectly categorized. This metric demonstrates the effectiveness of the classifier.
– Precision is the percentage of data points that were correctly classified relative to
those that were actually given a positive label.
– The percentage of correctly identified data points among all positive data points
is measured by recall.
– To evaluate the overall effectiveness of a binary classifier, the F1-score, a com-
bined measure of accuracy and recall, is used. The harmonic mean of recall and
precision can be used to calculate it.
Overall, the performance of a binary classifier can be evaluated using these measures
in order to choose the appropriate classifier for a given task.
confusion matrix, which includes information on the actual classifications and predic-
tions made by the model, can be used to calculate the accuracy of the model as well
as the error rate [16]. To determine accuracy, divide the total number of examples
that were correctly categorized (represented by the diagonal of the confusion matrix)
by the overall number of instances. The proportion or percentage of things that were
correctly categorized is shown below. Accuracy is subtracted from 1 to get error rate.
The percentage of incorrect classifications is shown here.
For binary classification problems, balanced accuracy is the most popular evaluation
statistic. When the definitions of a “true positive” and “true negative” are unclear,
this statistic, which is more reliable than standard accuracy, can be used to evaluate
the performance of a model [17]. By taking into consideration both FPs and FNs, a
model’s performance can be evaluated more accurately. Balanced accuracy can also
help to prevent a model from learning an imbalanced dataset and instead, ensure an
even level of performance across different types of predictions. This is especially im-
portant in situations involving important decisions where missed predictions can lead
to significant negative implications.
6.6 Cross-validation
A model validation technique called cross-validation evaluates how well the findings of
a statistical analysis transfer to a different dataset. It is primarily utilized in situations
where comparing machine learning algorithms is the major objective. The original sam-
ple is partitioned into k equal-sized subsamples at random for k-fold cross-validation.
K1 subsamples are utilized as training data out of a total of k subsamples, while the re-
maining k subsamples are used as validation data to test the model. Following that, the
6 Evaluation of AI model performance 133
cross-validation process is carried out k times, using the validation data from each of
the k subsamples just once [19]. By averaging the k outcomes from the folds, an estima-
tion can then be made. This approach has the advantage of repeated random subsam-
pling in that all observations are used for training and validation, and each observation
is used for validation exactly once.
Model selection and hyperparameter tuning are two essential parts of any machine
learning task. Model selection involves choosing a suitable learning algorithm, and
6 Evaluation of AI model performance 135
Model selection and hyperparameter tuning are vital processes for the optimization
of the machine learning process. As such, it should be an integral part of any machine
learning project.
By partitioning the data, training the model on a piece of the data, and then assessing
the model’s performance on a different subset of the data, the machine learning ap-
proach of cross-validation is used to evaluate a model. Since it divides the data into
“folds” and utilizes each data partition to estimate the performance of the model, this
technique is also used to identify the model with the highest accuracy [24]. It is crucial
for model selection since it not only aids in choosing the best model but also in identi-
fying the perfect set of hyperparameters, which enhances the model’s generalizability
and accuracy. Cross-validation aids in determining a model’s potential performance
when used with “unseen” data.
parameters, but instead selects a few at random to focus on. As new hyperparameter
sets are sampled, models are trained and evaluated, and the best (or best-performing)
hyperparameter set is selected [25]. In short, grid search provides an exhaustive
search of the hyperparameter space to find the best combination of parameters,
while random search provides a more stochastic approach, sampling a random com-
bination of parameters to find the best combination.
Addressing bias and fairness issues in model evaluation is an important step in build-
ing an effective and reliable model. It is essential to evaluate the model for potential
bias and fairness as it can lead to incorrect conclusions and lead to costly errors in
6 Evaluation of AI model performance 137
the future. In order to evaluate a model for bias and fairness, it is important to ana-
lyze how it treats different segments of the population. The evaluation should con-
sider statistical evidence such as correlations, measures of socio-economic disparities,
and other external factors that may influence predictions. Additionally, it is important
to review the model’s design, data, and results to identify potential bias or unfairness
[28]. Finally, the model should be tested in a variety of scenarios and circumstances in
order to ensure its accuracy. Addressing these issues will help ensure the model is
reliable and will lead to better decisions and outcomes.
Methods used to increase the human interpretability of machine learning models are
referred to as interpretability methods. These strategies include creating individual
conditional expectation (ICE) curves and partial dependence plots (PDPs), calculating
feature importance scores to determine the model’s most crucial characteristics, and
employing feature selection techniques to limit the amount of input features.
Evaluating interpretability methods can include assessing the utility of the meth-
ods for model developers and model users. For model developers, the effectiveness
6 Evaluation of AI model performance 139
and accuracy of the interpretability methods in helping to explain the model and de-
signing strategies to improve the model should be evaluated. For model users, the de-
gree to which interpretability methods enable them to understand and use the model
to make decisions should be assessed [33]. Additionally, the user experience associ-
ated with using these interpretability methods should be evaluated in terms of user
interface design, user friendliness, and the ease with which users can interpret the
outputs.
Best Practices:
– Explainability: Developers must ensure AI systems are explainable and easily un-
derstood by users. AI decision-making should be traceable, so users can track
why specific decisions were made.
– Security and privacy: Develop robust security measures to protect AI systems and
associated data. This includes creating safeguards against malicious actors, such
as ensuring users have proper authentication and authorization to access
systems.
– Data quality and Integrity: The reliability and integrity of the data have a big im-
pact on how accurate AI systems are. Ensure that data is dependable, consistent,
and frequently updated to maximize the accuracy of outcomes.
important to make sure that training data does not contain data that is not available
during inference.
A form of bias known as selection bias happens when a researcher decides to only
look at a portion of the population under consideration. This means that the sample
data they use could be disproportionately biased toward certain individuals or fac-
tors, resulting in an inaccurate depiction of the overall population [35].
Imbalance in evaluation datasets can occur when the evaluation dataset does not
accurately reflect the proportion of different class labels or categories in the original
dataset. For example, a large dataset may have 80% of samples from one class label,
while the evaluation dataset only contains 30% of those samples. This can lead to a
biased evaluation as the test model might not be able to accurately represent the full
range of data and experiences that it should be able to.
Any machine learning modeling project must include a thorough model evaluation. A
robust model evaluation involves a careful assessment of the model’s performance
against different types of data sets, different evaluation metrics, and under various
conditions [36]. Best practices for robust model evaluation include:
– Test on held-out data: Whenever possible, validate models on a set of held-out
data. This will give a more accurate assessment of how the model will perform in
production.
– Use multiple metrics: Utilizing numerous evaluation metrics can help you gain a
deeper knowledge of the model’s performance – using F1-score, accuracy, preci-
sion, recall, etc.
– Compare to baselines: Comparing the model’s performance against various base-
lines, such as random guessing or a straightforward model like logistic regression,
might be useful.
– Use cross-validation and hyperparameter tuning: Make that the model is not
under- or overfitting the data. Optimize the model using hyperparameter tweak-
ing, and evaluate the model’s generalizability using cross-validation.
– Check assumptions: Ensure that the assumptions that you are making about the
data are valid. This includes checking for data leakage or bias in the data.
142 Syed Ibad Ali et al.
References
[1] Allen, B., Dreyer, K., Stibolt Jr, R., Agarwal, S., Coombs, L., Treml, C., . . . & Wald, C. (2021). Evaluation
and real-world performance monitoring of artificial intelligence models in clinical practice: try it, buy
it, check it. Journal of the American College of Radiology, 18(11), 1489–1496.
[2] Tabbussum, R., & Dar, A. Q. (2021). Performance evaluation of artificial intelligence paradigms –
artificial neural networks, fuzzy logic, and adaptive neuro-fuzzy inference system for flood
prediction. Environmental Science and Pollution Research, 28(20), 25265–25282.
[3] Hicks, S. A., Strümke, I., Thambawita, V., Hammou, M., Riegler, M. A., Halvorsen, P., & Parasa,
S. (2022). On evaluation metrics for medical applications of artificial intelligence. Scientific Reports,
12(1), 5979.
[4] Li, L., Qin, L., Xu, Z., Yin, Y., Wang, X., Kong, B., . . . & Xia, J. (2020). Using artificial intelligence to
detect COVID-19 and community-acquired pneumonia based on pulmonary CT: evaluation of the
diagnostic accuracy. Radiology, 296(2), E65–E71.
[5] Dumitrascu, O., Dumitrascu, M., & Dobrotǎ, D. (2020). Performance evaluation for a sustainable
supply chain management system in the automotive industry using artificial intelligence. Processes,
8(11), 1384.
[6] Ghahramani, M., Qiao, Y., Zhou, M. C., O’Hagan, A., & Sweeney, J. (2020). AI-based modeling and
data-driven evaluation for smart manufacturing processes. IEEE/CAA Journal of Automatica Sinica,
7(4), 1026–1037.
[7] Khanagar, S. B., Al-Ehaideb, A., Maganur, P. C., Vishwanathaiah, S., Patil, S., Baeshen, H. A., . . . &
Bhandi, S. (2021). Developments, application, and performance of artificial intelligence in
dentistry–A systematic review. Journal of Dental Sciences, 16(1), 508–522.
[8] Wu, E., Wu, K., Daneshjou, R., Ouyang, D., Ho, D. E., & Zou, J. (2021). How medical AI devices are
evaluated: limitations and recommendations from an analysis of FDA approvals. Nature Medicine,
27(4), 582–584.
[9] Amin, M. M., Cambria, E., & Schuller, B. W. (2023). Will affective computing emerge from foundation
models and general ai? A first evaluation on chatgpt. arXiv preprint arXiv:2303.03186.
[10] Forghani, R., Savadjiev, P., Chatterjee, A., Muthukrishnan, N., Reinhold, C., & Forghani, B. (2019).
Radiomics and artificial intelligence for biomarker and prediction model development in oncology.
Computational and Structural Biotechnology Journal, 17, 995.
[11] Tobler, P., Cyriac, J., Kovacs, B. K., Hofmann, V., Sexauer, R., Paciolla, F., . . . & Hirschmann, A. (2021).
AI-based detection and classification of distal radius fractures using low-effort data labeling:
evaluation of applicability and effect of training set size. European Radiology, 31, 6816–6824.
[12] AI-Atroshi, C., Rene Beulah, J., Singamaneni, K. K., Pretty Diana Cyril, C., Neelakandan, S., &
Velmurugan, S. (2022). Automated speech based evaluation of mild cognitive impairment and
Alzheimer’s disease detection using with deep belief network model. International Journal of
Healthcare Management, 2022, 1–11.
[13] Gunning, D., & Aha, D. (2019). DARPA’s explainable artificial intelligence (XAI) program. AI Magazine,
40(2), 44–58.
[14] Zou, Y., Zheng, C., Alzahrani, A. M., Ahmad, W., Ahmad, A., Mohamed, A. M., . . . & Elattar, S. (2022).
Evaluation of artificial intelligence methods to estimate the compressive strength of geopolymers.
Gels, 8(5), 271.
[15] Yadav, S. P., & Yadav, S. (2020). Image fusion using hybrid methods in multimodality medical
images. Medical & Biological Engineering & Computing, 58, 669–687.
[16] Irfan, D., Tang, X., Narayan, V., Mall, P. K., Srivastava, S., & Saravanan, V. (2022). Prediction of Quality
Food Sale in Mart Using the AI-Based TOR Method. Journal of Food Quality, 2022, 1–13
6 Evaluation of AI model performance 143
[17] Yuvaraj, N., Srihari, K., Dhiman, G., Somasundaram, K., Sharma, A., Rajeskannan, S. M. G. S. M. A.,
. . . & Masud, M. (2021). Nature-inspired-based approach for automated cyberbullying classification
on multimedia social networking. Mathematical Problems in Engineering, 2021, 1–12.
[18] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). Energy‐efficient resource allocation
model for device‐to‐device communication in 5G wireless personal area networks. International
Journal of Communication Systems, 2023, 1–14.
[19] Yadav, S. P., Agrawal, K. K., Bhati, B. S., Al-Turjman, F., & Mostarda, L. (2022). Blockchain-based
cryptocurrency regulation: An overview. Computational Economics, 59(4), 1659–1675.
[20] Rajalakshmi, M., Saravanan, V., Arunprasad, V., Romero, C. T., Khalaf, O. I., & Karthik, C. (2022).
Machine Learning for Modeling and Control of Industrial Clarifier Process. Intelligent Automation &
Soft Computing, 32(1), 339–359.
[21] Manikandan, V., & Yuvaraj, N. (2019, May). Advanced expert system using particle swarm
optimization based adaptive network based fuzzy inference system to diagnose the physical
constitution of human body. In Emerging Technologies in Computer Engineering: Microservices in
Big Data Analytics: Second International Conference, ICETCE 2019, Jaipur, India, February 1–2, 2019,
Revised Selected Papers, 985, 349. Springer.
[22] Sangeetha, R., Logeshwaran, J., Rocher, J., & Lloret, J. (2023). An Improved Agro Deep Learning
Model for Detection of Panama Wilts Disease in Banana Leaves. AgriEngineering, 5(2), 660–679.
[23] Yadav, S. P., Zaidi, S., Mishra, A., & Yadav, V. (2022). Survey on machine learning in speech emotion
recognition and vision systems using a recurrent neural network (RNN). Archives of Computational
Methods in Engineering, 29(3), 1753–1770.
[24] Saravanan, V., & Neeraja, A. (2013, January). Security issues in computer networks and
stegnography. In 2013 7th International Conference on Intelligent Systems and Control (ISCO)
(pp. 363–366). IEEE.
[25] Gobinathan, B., Mukunthan, M. A., Surendran, S., Somasundaram, K., Moeed, S. A., Niranjan, P., . . .
& Sundramurthy, V. P. (2021). A novel method to solve real time security issues in software industry
using advanced cryptographic techniques. Scientific Programming, 2021, 1–9.
[26] Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model
for efficient resource allocation in 5G broadband communication networks. ICTACT Journal On Soft
Computing, 13(3), 2986–2994.
[27] Yadav, S. P., Mahato, D. P., & Linh, N. T. D. (Eds.). (2020). Distributed artificial intelligence: A modern
approach. Boca Raton: CRC Press.
[28] Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J., . . . &
Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey
wolf optimization in wireless sensor networks. Security and Communication Networks, 2022, 1–13
[29] Natarajan, Y., Kannan, S., & Mohanty, S. N. (2021). Survey of various statistical numerical and
machine learning ontological models on infectious disease ontology. Data Analytics in
Bioinformatics: A Machine Learning Perspective, 13(1), 431–442.
[30] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). L-RUBI: An efficient load-based
resource utilization algorithm for bi-partite scatternet in wireless personal area networks.
International Journal of Communication Systems, 2023, 1–15.
[31] Wan, Y., Tong, Y., Liu, Y., Huang, Y., Yao, G., Chen, D. Q., & Liu, B. (2022). Evaluation of the
combination of artificial intelligence and radiologist assessments to interpret malignant
architectural distortion on mammography. Frontiers in Oncology, 12, 1739.
[32] Hosamo, H. H., Nielsen, H. K., Kraniotis, D., Svennevig, P. R., & Svidt, K. (2023). Digital Twin
framework for automated fault source detection and prediction for comfort performance evaluation
of existing non-residential Norwegian buildings. Energy and Buildings, 281, 112732.
144 Syed Ibad Ali et al.
[33] Lee, D. H., Woo, S. E., Jung, M. W., & Heo, T. Y. (2022). Evaluation of odor prediction model
performance and variable importance according to various missing imputation methods. Applied
Sciences, 12(6), 2826.
[34] Markus, A. F., Kors, J. A., & Rijnbeek, P. R. (2021). The role of explainability in creating trustworthy
artificial intelligence for health care: A comprehensive survey of the terminology, design choices,
and evaluation strategies. Journal of Biomedical Informatics, 113, 103655.
[35] Orhan, K., Bayrakdar, I. S., Ezhov, M., Kravtsov, A., & Özyürek, T. A. H. A. (2020). Evaluation of
artificial intelligence for detecting periapical pathosis on cone‐beam computed tomography scans.
International Endodontic Journal, 53(5), 680–689.
[36] Schwartz, R., Dodge, J., Smith, N. A., & Etzioni, O. (2020). Green ai. Communications of the ACM,
63(12), 54–63.
K. Arthi, Veeramalai Sankaradass, Nikhat Parveen, Muralidharan J.
7 Methods of cross-validation
and bootstrapping
Abstract: A statistical method called cross-validation is used to determine a machine
learning model’s parameters and gauge its correctness. It is a type of resampling
method where the dataset is arbitrarily split into training and test sets. Using different
subsets of the same data, the model is tested and trained. Comparing the projected
values from the testing data with the actual values of the same data allows one to
evaluate the model’s performance.
7.1 Introduction
A statistical resampling method called “bootstrapping” is used to evaluate the effec-
tiveness of a machine learning model. This technique replaces the need for k-fold
cross-validation and can quickly provide measures of accuracy on unseen data. Boot-
strapping randomly generates new datasets out of the original dataset by using sam-
pling with replacement. This technique is an efficient way of obtaining new datasets
with a higher degree of variance than the original dataset [1]. A machine learning
model’s performance can be evaluated more robustly and precisely with the use of
bootstrapping.
https://doi.org/10.1515/9783111323749-007
146 K. Arthi et al.
k-Fold cross-validation is one technique for evaluating a machine learning model’s ac-
curacy. With this method, a dataset is split into k subgroups, and k-rounds of training
and testing are performed on each subset. Each round uses a different subset of the
data for training and the remaining data for testing. Every round includes an evalua-
tion of the model’s accuracy. The mean accuracy of the k trials is then used to estimate
the model’s accuracy. k-Fold cross-validation is an effective method for assessing a mod-
el’s accuracy because it not only guards against overfitting but also has the capacity to
produce more accurate estimations of the model accuracies.
7 Methods of cross-validation and bootstrapping 147
The k-fold cross-validation method is used to assess the accuracy and fairness of a ma-
chine learning algorithm. k-Fold cross-validation randomly selects k subsets (or “folds”)
from a dataset [5]. Then, each subset is used as a training set and a test set once each.
After this technique has been applied k times, the final result is obtained by averaging
the outcomes of all k trials.
1. Data division into k subsets (folds): The first step is to divide the data into k
equal-sized, nonoverlapping sections (or “folds”).
2. Train and test the model k times: The model is then trained and tested k times,
each on a distinct fold.
3. Average the model results: To obtain the final model score, the outcomes from all
k-folds are averaged.
Multiple test iterations are conducted using the k-fold cross-validation technique to
verify the accuracy of a model. The training set for each iteration is a portion of the
data, and the testing set is the remainder. The average testing accuracy of different
models can be calculated by doing a loop through all the data.
Benefits include:
– A reduction in overfitting’s negative effects. k-Fold cross-validation gives an objective
assessment of the model’s performance because each data point is used for testing.
– It provides an estimated population error rate: By performing multiple iterations,
the average test accuracy can be estimated, thus providing a good indication of
the model’s performance on unseen data.
Limitations:
– Inefficient use of data: Since each iteration requires a portion of the data to be
used for training and a portion for testing, it is not possible to use all the data
simultaneously.
– It may suffer from selection bias: Selecting a randomly chosen subset of data
points from the pool may not be representative of the population. Thus, k-fold
cross-validation may lead to inaccurate results.
Stratified cross-validation is a method of model validation, which helps ensure that all
the data in a given set of samples are equally represented in the different validation
148 K. Arthi et al.
sets. It is typically used when there is a significant imbalance in the data, such as when
classes are unevenly distributed. Stratified cross-validation is a type of cross-validation
that attempts to address this problem by dividing the data into subsets according to the
class labels [6]. Then, it partitions the subsets into testing and validation sets. As a re-
sult, each sample is guaranteed to be represented in both the validation and test sets.
This strategy aids in ensuring that a model is reflective of the entire dataset, rather
than just the classes with the highest percentages.
For evaluating and contrasting several models or approaches, machine learning em-
ploys a sort of cross-validation called stratified cross-validation. By dividing the data
into various folds, it becomes easier to assess how well machine learning models are
performing. With stratified cross-validation, each class is represented equally in both
the training and test sets for each fold [7]. As a result, the model will be tested and
trained using data sets that are representative of the entire population. Additionally,
when working with tiny datasets, it can aid in preventing bias or distorted outcomes.
In handling imbalanced datasets, application is a set of tools and techniques that can
be used to deal with the problem of imbalanced datasets. These tools and techniques
can help balance the data by either adding more minority class examples or by re-
moving majority class examples. Some of these techniques include downsampling,
clustering, and oversampling. When a dataset is downsampled, instances of the ma-
jority class are randomly removed while remaining examples of the minority class
are retained. Clustering algorithms can be used to assign similar examples into clus-
ters and remove outliers [8]. Oversampling involves creating additional instances of
the minority class from existing examples. The application of these tools and techni-
ques can help reduce the bias in machine learning models trained on imbalanced
datasets.
testing sets [9]. The accuracy of the model is then assessed by averaging the outcomes
of each iteration. Particularly when the data set is small, this method can offer esti-
mates of model accuracy that are more accurate than those provided by conventional
cross-validation techniques.
LOOCV is a validation technique that involves removing one data point from the dataset
and using the remaining details to develop a predictive model. After that, the model is
assessed using the same criteria as the remaining data on the deleted data point. Every
last data point in the set is subjected to this process once more, and the results are aver-
aged to provide an overall assessment of model performance [10]. When determining
the precision and dependability of predictive models, LOOCV is a popular method since
it is effective at decreasing bias and volatility in model evaluation.
Cross-validation that uses one data point for testing while using the remaining data
points for training is known as leave-one-out cross-validation (LOOCV). When N is the
total number of data points, this operation is performed N times. Since the complete
dataset must be examined several times, this is one of the most demanding types of
validation and is frequently quite time-consuming.
The implications of using LOOCV are that the accuracy of the model is maximized
while minimizing the bias. Since all observations are used for both test and training,
the risk of overfitting or underfitting is significantly reduced. This method yields
more reliable predictions than other cross-validation techniques [11].
Nested cross-validation is a technique used to evaluate the model selection and parame-
ter tuning of a predictive model. It involves conducting an inner loop of cross-validation
to tune the model and an outer loop of cross-validation to test the effectiveness of the
model selection and/or parameter tuning. This technique is used to minimize overfitting
of the model and to ensure an unbiased estimate of the model’s generalization error [12].
This technique is more computationally expensive than traditional test/train splitting, but
can give higher accuracy estimated out-of-sample performance.
150 K. Arthi et al.
When working with time series data, a cross-validation approach known as time se-
ries cross-validation is employed. The process is splitting the data into two or more
separate, nonoverlapping subgroups, training a model on one of those subsets, and
testing it on the other. The model can learn temporal information through this type of
cross-validation, which also evaluates the model’s capacity for generalization across
time. It is especially helpful when working with lengthy time series, as creating more
subsets for traditional cross-validation methods like k-fold cross-validation can take a
lot of time [14]. Additionally, time series cross-validation can be used to assess the im-
pact of using a different window size for training or testing, as well as the impact of
using various training and testing datasets.
multiple training and test sets. The model is trained on the training sets and evaluated
against the test sets. This method effectively prevents the model from overfitting to
any single point in time, which might not represent other points in time. It also ena-
bles the model to generalize knowledge about the time series and be better prepared
for future forecasts [15]. To evaluate the generalization of forecasting models, calibra-
tion metrics, and feature relevance, time series cross-validation can be used.
Rolling window approach is a type of time series analysis used to estimate a value or
calculate metrics based on a limited-size window. A rolling window comprises a set of
consecutive time periods and the most recent data points are used to estimate a value
or calculate a metric. This approach captures recent inputs and creates dynamic time
frames for analysis. In contrast, the expanding window approach works with the en-
tire data set instead of just the recent time period data [16]. This is a useful technique
when the data set is composed mostly of historical observations and you are trying to
map out logic or detect trends. The expanding window approach looks at increasing
amounts of data, instead of just the most recent, to create a wider frame of reference.
Additionally, this technique can capture changes in the underlying trend and identify
shift points in the data.
7.3 Bootstrapping
Using replacement sampling from the original sample, the approach known as “boot-
strapping” can be used to estimate the sampling distribution of an estimator. This
method is most frequently used to calculate standard errors and confidence intervals
for population parameters. This method can do away with the requirement for diffi-
cult mathematical calculations that are required to derive the sampling distribution
of an estimate. The procedure involves taking several samples from the original data-
set with replacement and creating the sampling distribution by applying the estimator
of interest to each sample. This distribution can provide insight into the variability of
the estimator [17]. Bootstrapping can be used for a variety of different population pa-
rameters such as means, medians, variances, correlations, and regression coefficients.
Additionally, it can also be used to test hypotheses, compare differences between
groups, and construct confidence intervals.
152 K. Arthi et al.
With replacement, a number of random samples are picked from a larger population
as part of the model evaluation approach known as “bootstrapping.” By analyzing the
variability of a predictive model’s performance across resampled datasets, it is possi-
ble to evaluate a predictive model’s stability and reliability. Bootstrapping can be
used to assess data, estimate uncertainty, and validate models. It is also used to calcu-
late the standard deviation of a statistical model’s parameters [20]. Bootstrapping is
often used in conjunction with cross-validation to provide good estimates on model
7 Methods of cross-validation and bootstrapping 153
By combining the predictions of several weak models, such as decision trees, boot-
strap aggregating, also known as bagging, is a machine learning approach used to in-
crease the generalization accuracy of a given model. Rather than use the same set of
data to fit each tree, the technique uses a bootstrapping technique where randomly
chosen samples are used for training each tree.
The benefit of bootstrap aggregating over a single decision tree model is that it re-
duces the variance and increases the predictive accuracy by taking the average of mul-
tiple trees. Bagging also helps reduce the risk of overfitting by creating an ensemble of
random forests that ultimately creates a more stable prediction. Additionally, it can de-
crease the time needed to build a model, as individual trees can be built in parallel.
A bagging model, a type of ensemble learning methodology, can be evaluated for accu-
racy using out-of-bag (OOB) estimates. OOB estimate is founded on a theory known as
bootstrap aggregation, or “bagging” for short. The goal of bagging is to build numerous
iterations of a single model, each using a different set of training data, and then aggre-
gate the output of each model to produce a more reliable model. By examining the OOB
samples, OOB estimation measures the correctness of each individual model [21]. The
OOB samples are those portions of the training dataset that were not used to train the
model. The correctness of OOB samples for each model is monitored. The average of the
accuracy of each individual model makes up the bagging model’s overall accuracy. Any
bagging model, including random forests, support vector machines, and neural net-
works, can be evaluated for accuracy using OOB estimation.
The main difference between the two terms is in their underlying principles. The
term analytics has traditionally referred to the use of data to analyze trends and to
inform decision-making. It encompasses the use of methods such as statistics, data
mining, machine learning, and predictive analysis to understand, describe, and pre-
dict user behavior. It is typically used to support decision making in areas such as
product marketing and development.
7 Methods of cross-validation and bootstrapping 155
The goal of artificial intelligence (AI), on the other hand, is to make computers intelli-
gent by developing systems that can learn and solve problems. It uses techniques such as
neural networks, natural language processing, and Bayesian networks to create systems
that are able to reason and react in complex environments. AI is focused on replicating
certain cognitive abilities, enabling the development of more intelligent computer sys-
tems, which can engage in tasks that would require higher level cognitive skills from a
human. AI is seen in applications such as self-driving cars, robotics, and facial recognition
[23]. The analytics is focused on the use of data to analyze trends and inform decision-
making, while AI utilizes techniques to create intelligent computer systems to solve com-
plex problems.
The output of a machine learning model can be verified via repeated cross-validation
(RCV). The training set and test set are created by randomly splitting the dataset. The
model is then trained on the training set using the specified algorithm. The test set is
then used to evaluate the trained model’s performance after splitting the dataset at
random, training, and evaluating the model, and repeating the process many times
with different random splitting values. This enables us to accurately assess the perfor-
mance of the model by taking into account the variability due to different splits. RCV
is commonly used to tune the hyperparameters of a model, such as regularization in
logistic regression or learning rate in neural networks.
RCV is a statistical method for validly assessing predictive models. Iteratively, a given da-
taset is divided into training and test sets at random in this method. The predictive model
is then tested on the test dataset after being tested on the training dataset. Then, the pro-
cedure is repeated a predetermined number of times with a different division of the test
and training data each time. The performance of the model is then calculated by averag-
ing the results of all iterations. Since it accounts for the variability in the training and test
data used to evaluate the model, this method’s key advantage is that it enables estimates
of the model’s performance to be more precise. Additionally, by using this method, any
faults with the modeling procedure itself, such as overfitting difficulties, can be found.
The basic k-fold technique is extended to account for the relative class distribution of
the data, and this is known as stratified group k-fold cross-validation. The data is di-
vided into k groups for this type of validation, and each group is stratified to make
sure that the distribution of the various classes within each group is the same as it
was in the original data set. As a result, the validation findings are more accurate rep-
resentations of the complete data set. When performing stratified group k-fold cross-
validation, for instance, each group should have 30% class A and 70% class B for a
classification problem where the data set is composed of 30% class A and 70% class B.
folds each containing an equal amount of data from all groups (or strata) in the data-
set. In this technique, the data remains grouped together while being randomly parti-
tioned into the folds. This is useful when the dataset contains related data that should
not be separated, such as when working with grouped data containing different age
ranges or different ethnicities [25]. The stratified approach ensures that each fold con-
tains balanced amounts of different groups from the dataset, in order to allow model
evaluation to be as accurate as possible. After all the folds have been generated, each
fold can be used in a separate training and validation process, with the average per-
formance across all folds being used as the final evaluation result.
The Monte Carlo cross-validation (MCCV) technique uses a dataset that has been ran-
domly divided into training and test sets, with the training set being used to fit a
model and the test set being used to assess the model’s accuracy. This type of cross-
validation is referred to as a “Monte Carlo” method because it randomly samples
from the data before partitioning it. This allows the modeler to use all available data
when constructing the model, without needing to split it beforehand. Then, repeat-
edly, subsets of the data are randomly sampled and a model is trained on one subset
and evaluated on another. The average performance over all the iterations is then
used as the final accuracy of the model.
MCCV and random sampling are used to assess the correctness of a model. Using this
technique, the dataset is randomly split into training and testing sets. The testing set
is used to evaluate the model once it has been created using the training set.
The training and testing sets are then divided into several subsets using MCCV
[26]. Each of these subsets is then used in combination with the rest to evaluate the
model. In this manner, the model’s performance is tested multiple times on different
combinations of data, providing a more accurate assessment of its performance.
Random sampling with MCCV is a powerful tool for assessing model accuracy, as
it allows for an efficient assessment of both overfitting and underfitting. However, it
is important to note that the results from MCCV should be taken with a grain of salt
and should always be used alongside other assessment methods when evaluating the
accuracy of a model.
158 K. Arthi et al.
Time series bootstrapping is a method of resampling that uses the past data points of
a time series to simulate new data points. This method is used by data scientists to
explore the variance and predict future trends in the data. Bootstrap samples are gen-
erated from the series of past data points by selecting random data points with re-
placement and are then used in various statistical tests [27]. This method is especially
useful when the data set is too small to provide accurate or reliable results. By gener-
ating a bootstrap sample from the original data, the variance and potential trends can
be explored, and realistic predictions can be made.
A method known as extending bootstrapping to time series data makes use of resam-
pling to reveal patterns in the behavior of a time series while taking into consider-
ation its complexity. By selecting replacement samples at random from a data set, the
resampling approach known as “bootstrapping” can estimate the distribution of a sta-
tistic. It is frequently used to replicate the uncertainty surrounding a statistic, allow-
ing us to draw conclusions about the population’s unknowable values.
When applied to time series data, it creates simulated time series that reflect the
underlying distribution of the actual time series to be tested [28]. It can be used to
assess the reliability of an estimate, identify potential outliers, provide insights into
the structure of the time series, and allow comparison between multiple time series.
By simulating future values of a time series, it can also be used to predict future val-
ues and analyze their stability.
Block bootstrap and the stationary bootstrap are two methods of sampling data that can
be used to create a larger set of data for analysis and prediction. Block bootstrap is a
method of sampling data that breaks the original dataset into smaller blocks called boot-
strap replicates. This allows for the creation of a larger sample from the original dataset.
The advantage of this methodology is that the original dataset tends to be preserved in
the blocks, since each block will contain some of the original $\mu$. variability.
The stationary bootstrap is a sampling technique that preserves the structure of
the original data. This is done by sampling random points in the data, and assigning
them to a new sample set. When the underlying generation method of the data is suf-
ficiently varied and applicable to a wide range of data, it is valuable. It permits the
generation of numerous distinct samples from a single initial dataset while maintain-
ing the original $mu$ variability.
7 Methods of cross-validation and bootstrapping 159
The accuracy of a predictive model can be increased by using cross validation, which
is a crucial component of model validation. We may ensure that the model is properly
evaluated and that the best outcomes are produced by taking into account the practi-
cal concerns and best practices listed above.
Managing small datasets with few samples can be difficult. It is usually more challeng-
ing to work with small datasets with few samples than with larger datasets with
many samples. This is because the smaller sample size does not provide as much in-
formation as the larger sample size, making it hard to accurately analyze or make
valid predictions based on the limited data [30]. To make the best use of the limited
information, techniques such as feature engineering, data augmentation, and cross-
validation can be used to generate more meaningful results. Additionally, sampling
techniques such as bootstrapping can be used to create more robust datasets. Finally,
160 K. Arthi et al.
machine learning algorithms that are designed to work well with small datasets
should be chosen in order to maximize the predictive power of the limited data.
Selecting the appropriate evaluation metrics is a vital stage in any machine learning
process. Depending on the problem, different evaluation measures might be selected.
For example, if the task at hand is a classification task, some common metrics that
can be used include accuracy, precision, recall, F1 score, and area under the curve.
When a regression task is involved, the preferred metrics are frequently the mean
absolute error, mean squared error, and R-squared values [31]. Additionally, depend-
ing on the difficulty of the task, certain metrics may be required, including sensitivity,
specificity, confusion matrix, true positive, false positive, and false negative. The cho-
sen metrics must accurately reflect the quality of the output of the system being built,
regardless of how the challenge may affect them.
The proper reporting and interpretation of results is a critical step in any research
project. Proper reporting includes providing an accurate and complete description of
the study, presenting the results in an appropriate format and fashion, and discussing
7 Methods of cross-validation and bootstrapping 161
Limitations of bootstrapping:
– Bootstrapping is limited to the size of the existing data set. It is also subject to
strong assumptions, such as independence and identical distribution, which can
lead to inaccuracies when the assumptions are violated.
some limitations and pitfalls. Cross-validation can be sensitive to the choice of sample
splits, so extreme care must be taken to avoid sample splitting bias. If the process of
splitting the data set is not carefully planned and executed, it is possible to end up with
an easy-to-fit model, which will produce unrealistic results when applied on unseen
data [32]. The bootstrapping approach can potentially be hampered by overestimating a
method’s capacity for generalization to fresh data. Using various resampling methods,
such as subsampling or leave-one-out cross-validation, can help to alleviate this. Addi-
tionally, some characteristics of the data used to produce the sample may have an im-
pact on how the data are manipulated, and the manipulated data may not accurately
represent the population from which the sample was collected. Despite the fact that
cross-validation and bootstrapping can be helpful in assessing a model’s performance,
it is crucial to be aware of their drawbacks and danger zones. When using these techni-
ques, careful thought should be applied to guarantee appropriate findings.
The emergence of resampling techniques has been a major part of greater advance-
ments in predictive modeling, machine learning, and research in general. Resampling
techniques provide data scientists with the ability to effectively measure the perfor-
mance of a given model, by evaluating the accuracy of the model on data which has
been repeatedly sampled. Through these techniques, researchers are able to find out
if a given model is too complex or too simple. This is done by changing data and fea-
ture selection and resampling to create new training sets. Additionally, these techni-
ques can help debug models for unexplained results, giving researchers greater
confidence in their models. Furthermore, by providing new ways of testing the perfor-
mance of a model, resampling techniques can help researchers find areas of opportu-
nity and potential improvements. Ultimately, these techniques can help accelerate the
development and deployment of more reliable algorithms in data science research.
The goal of model evaluation methods is to ensure models are properly trained and
evaluated on data that closely follows the real-world environment they will be de-
ployed in. As machine learning models become more pervasive, the need for more
advanced evaluation methods that can analyze more complex datasets and capture
subtle characteristics of datasets is growing. In the future, these model evaluation
methods may incorporate more techniques that use adversarial or generative model-
ing to better mimic data from the real world. Additionally, tools that are better able to
understand and detect outliers, noisy data, and patterns in datasets may improve
7 Methods of cross-validation and bootstrapping 163
References
[1] Brodeur, Z. P., Herman, J. D., & Steinschneider, S. (2020). Bootstrap aggregation and cross‐validation
methods to reduce overfitting in reservoir control policy search. Water Resources Research, 56(8),
e2020WR027184.
[2] Rafało, M. (2022). Cross validation methods: Analysis based on diagnostics of thyroid cancer
metastasis. ICT Express, 8(2), 183–188.
[3] Ramesh, G., Logeshwaran, J., & Kumar, A. P. (2023). The smart network management automation
algorithm for administration of reliable 5G communication networks. Wireless Communications and
Mobile Computing, 2023, 1–12.
[4] Yadav, S. P., & Yadav, S. (2020). Image fusion using hybrid methods in multimodality medical
images. Medical & Biological Engineering & Computing, 58, 669–687.
[5] Vrigazova, B. (2021). The proportion for splitting data into training and test set for the bootstrap in
classification problems. Business Systems Research: International Journal of the Society for
Advancing Innovation and Research in Economy, 12(1), 228–242.
[6] Vrigazova, B., & Ivanov, I. (2020). Tenfold bootstrap procedure for support vector machines.
Computer Science, 21, 241–257.
[7] de Rooij, M., & Weeda, W. (2020). Cross-validation: A method every psychologist should know.
Advances in Methods and Practices in Psychological Science, 3(2), 248–263.
[8] Uyun, S., & Sulistyowati, E. (2020). Feature selection for multiple water quality status: Integrated
bootstrapping and SMOTE approach in imbalance classes. International Journal of Electrical and
Computer Engineering, 10(4), 4331.
[9] Irfan, D., Tang, X., Narayan, V., Mall, P. K., Srivastava, S., & Saravanan, V. (2022). Prediction of quality
food sale in mart using the AI-Based TOR method. Journal of Food Quality, 2022, 1–13.
[10] Yadav, S. P., Zaidi, S., Mishra, A., & Yadav, V. (2022). Survey on machine learning in speech emotion
recognition and vision systems using a recurrent neural network (RNN). Archives of Computational
Methods in Engineering, 29(3), 1753–1770.
[11] Yuvaraj, N., & Vivekanandan, P. (2013, February). An efficient SVM based tumor classification with
symmetry non-negative matrix factorization using gene expression data. In 2013 International
Conference on Information Communication and Embedded Systems (Icices) (pp. 761–768). IEEE.
164 K. Arthi et al.
[12] Fernandez-Felix, B. M., García-Esquinas, E., Muriel, A., Royuela, A., & Zamora, J. (2021). Bootstrap
internal validation command for predictive logistic regression models. The Stata Journal, 21(2),
498–509.
[13] Sivaram, M., Yuvaraj, D., Mohammed, A. S., Manikandan, V., Porkodi, V., & Yuvaraj, N. (2019).
Improved enhanced DBTMA with contention-aware admission control to improve the network
performance in MANETS. CMC-Computers Materials & Continua, 60(2), 435–454.
[14] Sangeetha, R., Logeshwaran, J., Rocher, J., & Lloret, J. (2023). An Improved Agro Deep Learning
Model for Detection of Panama Wilts Disease in Banana Leaves. AgriEngineering, 5(2), 660–679.
[15] Vrigazova, B. P. (2020). Detection of malignant and benign breast cancer using the Anova-
Bootstrap-SVM. Journal of Data and Information Science, 5(2), 62–75.
[16] Larracy, R., Phinyomark, A., & Scheme, E. (2021, November). Machine learning model validation for
early stage studies with small sample sizes. In 2021 43rd Annual International Conference of the
IEEE Engineering in Medicine & Biology Society (EMBC) (pp. 2314–2319). IEEE.
[17] Takada, T., Nijman, S., Denaxas, S., Snell, K. I., Uijl, A., Nguyen, T. L., . . . & Debray, T. P. (2021).
Internal-external cross-validation helped to evaluate the generalizability of prediction models in
large clustered datasets. Journal of Clinical Epidemiology, 137, 83–91.
[18] Cho, G., Jung, K., & Hwang, H. (2019). Out-of-bag prediction error: A cross validation index for
generalized structured component analysis. Multivariate Behavioral Research, 54(4), 505–513.
[19] Zhang, Y., Qiao, S., Ji, S., Han, N., Liu, D., & Zhou, J. (2019). Identification of DNA–protein binding
sites by bootstrap multiple convolutional neural networks on sequence information. Engineering
Applications of Artificial Intelligence, 79, 58–66.
[20] Harrou, F., Saidi, A., & Sun, Y. (2019). Wind power prediction using bootstrap aggregating trees
approach to enabling sustainable wind power integration in a smart grid. Energy Conversion and
Management, 201, 112077.
[21] Saravanan, V., & Sumathi, A. (2012). Handoff mobiles with low latency in heterogeneous networks
for seamless mobility: A survey and future directions. European Journal of Scientific Research, 81(3),
417–424.
[22] Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model
for efficient resource allocation in 5G broadband communication networks. ICTACT Journal On Soft
Computing, 13(3), 2986–2994.
[23] Yadav, S. P., & Yadav, S. (2019). Mathematical implementation of fusion of medical images in
continuous wavelet domain. Journal of Advanced Research in Dynamical and Control System, 10(10),
45–54.
[24] Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J., . . . &
Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey
wolf optimization in wireless sensor networks. Security and Communication Networks, 2022, 1–10.
[25] Kim, B., Xu, C., & Barber, R. (2020). Predictive inference is free with the jackknife+-after-bootstrap.
Advances in Neural Information Processing Systems, 33, 4138–4149.
[26] Khan, S., Saravanan, V. N. G. C., Lakshmi, T. J., Deb, N., & Othman, N. A. (2022). Privacy protection of
healthcare data over social networks using machine learning algorithms. Computational
Intelligence and Neuroscience, 2022, 1–9.
[27] Natarajan, Y., Kannan, S., & Mohanty, S. N. (2021). Survey of various statistical numerical and
machine learning ontological models on infectious disease ontology. Data Analytics in
Bioinformatics: A Machine Learning Perspective, 13(1), 431–442.
[28] Yuvaraj, N., Srihari, K., Dhiman, G., Somasundaram, K., Sharma, A., Rajeskannan, S. M. G. S. M. A., . . . &
Masud, M. (2021). Nature-inspired-based approach for automated cyberbullying classification on
multimedia social networking. Mathematical Problems in Engineering, 2021, 1–12.
[29] Yousefa, W. A. (2020). A leisurely look at versions and variants of the cross validation estimator. stat,
1050, 9.
7 Methods of cross-validation and bootstrapping 165
[30] Ho, S. Y., Phua, K., Wong, L., & Goh, W. W. B. (2020). Extensions of the external validation for
checking learned model interpretability and generalizability. Patterns, 1(8), 100129.
[31] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). Energy‐efficient resource allocation
model for device‐to‐device communication in 5G wireless personal area networks. International
Journal of Communication Systems, 2023, 1–14.
[32] Rani, P., Verma, S., Yadav, S. P., Rai, B. K., Naruka, M. S., & Kumar, D. (2022). Simulation of the
lightweight blockchain technique based on privacy and security for healthcare data for the cloud
system. International Journal of E-Health and Medical Communications (IJEHMC), 13(4), 1–15.
S. Ramasamy, Kantharaju H. C., N. Bindu Madhavi, Haripriya M. P.
8 Meta-learning through ensemble
approach: bagging, boosting, and random
forest strategies
Abstract: Meta-learning through ensemble approaches is an intriguing subfield of ma-
chine learning research. With this method, a more comprehensive learning model is
created by combining many machine learning methods, including neural networks
and support vector machines. By using an ensemble of models, meta-learning techni-
ques are able to produce more robust results than individual algorithms alone. In ad-
dition, ensemble techniques are advantageous because they can easily be expanded
to accommodate additional data sources or algorithms. This approach can also embed
more knowledge from the data into a more powerful meta-model, which allows the
system to generalize better and discover patterns more accurately. In short, meta-
learning through ensemble approaches is an effective and useful technique for tack-
ling challenging problems in machine learning.
https://doi.org/10.1515/9783111323749-008
168 S. Ramasamy et al.
meta-learning algorithms can swiftly and effectively learn how to adapt to problems
that have never been seen before. Meta-learning’s objective is to develop models that
are easily generalizable and adaptable to data from other areas [2].
Meta-learning is an important tool for data scientists and researchers as it in-
creases the efficiency and effectiveness of the whole machine learning process. In-
stead of having to spend time training a model from scratch each time a task needs to
be accomplished, the meta-learning model can be quickly configured to the new task
using the skills it has acquired from prior tasks. Recently, meta-learning has been
used successfully in fields like reinforcement learning, natural language processing,
and computer vision. Machine learning’s expanding topic of meta-learning has a wide
range of intriguing applications [3].
A sort of machine learning technique, known as ensemble learning, combines the pre-
dictions of several independent models to produce a single, improved model. This
method is applied to increase the model’s accuracy and robustness. There are five
standard ensemble learning methods:
– Boosting: Boosting builds up a strong classifier gradually, based on the weak clas-
sifiers or the simple base classifiers. It works by giving more weight to the
wrongly classified examples so that those examples are given more attention in
the next round of training.
– Bagging: A sort of ensemble learning technique, bagging creates a single predic-
tion by averaging the model outputs. It functions by using the same algorithm to
train numerous models, which are then combined to produce a forecast.
– Stacking: As an ensemble technique, stacking combines the outcomes of various
models into a single outcome. It functions by using the output of one model as
input for another model.
– Blending: Blending functions by averaging the results of various models’ fore-
casts. It is frequently used to lessen the variation in a model’s predictions.
– Voting: Voting is an ensemble learning technique that integrates the forecasts
from various models. Every model makes a prediction, and the final prediction is
determined by taking the vote of the majority of all the predictions.
170 S. Ramasamy et al.
8.3 Bagging
Bagging is a type of ensemble method in which the same classification algorithm is
used repeatedly on different subsets of the training data. This is an approach used to
increase the accuracy of a predictive model by reducing the variance in its predic-
tions. Bagging helps improve accuracy by combining the results of different classifiers
and reducing their variance [6]. This process averages out the errors of individual
classifiers and thus reduces the overall effects of noise. Bagging can be used with any
kind of predictive model, although it is most frequently used with decision trees since
it helps lessen the instability of a single decision tree.
The process known as bootstrap aggregating, often referred to as bagging, reduces the
variance of individual estimates by merging them, producing a more accurate aggre-
gate estimate. This is done by creating multiple bootstrapped versions of the original
data set and then combining the different estimates to create an aggregated value.
Bagging can also help control overfitting in machine learning applications. The bag-
ging algorithm works by randomly selecting from the available data points and then
combining the independent estimations that are returned. This reduces the variance
of the overall estimate by averaging out the individual estimates.
grated, typically using an average. This generally results in more accurate forecasts
than a single model, and helps to lower the variance of the predictions.
Bagging helps improve the learning algorithm’s performance in a number of
ways. Firstly, it increases the generalization of the model, meaning that the model can
more accurately predict values for unseen data based on the data it has learned from.
Secondly, it helps reduce overfitting, since it shows the model different training sets
and therefore leads to more robust models [9]. Finally, it defends against data distor-
tion, since the different training sets randomized to be created during the bagging
process make it more difficult for any given batch of data to overly skew the model’s
predictions.
8.4 Boosting
Boosting is a type of machine learning algorithm that iteratively learns from data, al-
lowing models to remain flexible and produce more accurate results. It does this by
combining a set of weak learners (e.g., shallow decision trees) and creating a strong
learner. Weak learners are instructed in stages, with each one attempting to correct
the one before it. The final prediction is then created by adding up each individual’s
predictions using a weighted majority vote (or sum).
Boosting is an ensemble learning strategy where several weak learners are combined
to form one strong learner. It operates by building a progressively additive model,
where each new model is built using the series of residuals from the prior one. The
boosting process consists of adding each successive model to the ensemble to reduce
the overall error of the final model [10]. Each successive model is an attempt to cor-
rect the errors of the previous model. The objective is to produce a strong learner that
will improve the predictive performance on unseen data. Boosting algorithms have
the advantage of not making strong assumptions about the data shape, and being able
to absorb bias from previous models.
An ensemble learning technique called AdaBoost, short for Adaptive Boosting, is used
to create predictive models. It operates by integrating several weak learners into one
strong learner and is trained iteratively. Then, predictions are made by using this
powerful learner. It functions by initially creating a weak classifier using a subset of
172 S. Ramasamy et al.
data. This classifier is employed to predict outcomes based on the provided data [11].
If it makes mistakes, the algorithm increases the weight of that instance and tries to
create a better classifier. The process is repeated until a strong learner is formed. The
AdaBoost algorithm is often used in conjunction model with decision trees, which pro-
vide the weak learners used by the algorithm. It is a popular choice of algorithm due
to its simplicity, effectiveness, and robustness.
In order to generate an ensemble strong learner model, the boosting machine learning
method combines a number of weak learners, each of which concentrates on identify-
ing a certain feature of the dataset. It works by training the weak learners sequentially,
each time refocusing on the samples that were misclassified by the previous learner. At
the end of the process, the ensemble is provided with a majority vote on the class label
assigned to a given sample [13]. Many sophisticated learning algorithms, including
XGBoost, are built using this technique of integrating multiple weak predictive models
to produce a powerful statistical model.
Advantages of boosting
– It is a sequential learning technique that uses previous versions of the model to
improve upon itself.
8 Meta-learning through ensemble approach 173
– Boosting algorithms work particularly well with noisy data due to the use of
weak learners.
– They are able to handle a large variety of data types.
– They are relatively insensitive to model parameter specifications.
– They allow for the definition of complex relationships between variables.
– Boosting algorithms come with fewer parameters to tune compared to deep learn-
ing algorithms.
Object detection, anomaly detection, online advertising, credit scoring, fraud detec-
tion, natural language processing (NLP), and recommendation systems are just a few
of the real-world applications that make use of boosting techniques. Boosting algo-
rithms are effective techniques for enhancing the dependability and accuracy of ma-
chine learning models. They are used in many applications to find patterns in data
and generate predictions that are more accurate and reliable than those produced by
conventional techniques. Boosting algorithms can also be used to reduce bias and var-
iance in a model’s predictions. For example, they may be used to identify correlations
in data that would otherwise be overlooked in traditional models. Boosting algorithms
can also be used to improve accuracy when classifying data and reduce the likelihood
of the occurrences of false positives or false negatives [14]. The ability to identify pat-
terns in data is especially useful for medical diagnosis, fraud detection, and anomaly
detection. Additionally, boosting techniques are employed in NLP applications to en-
hance the precision and speed of word, phrase, and syntax recognition. Finally, online
advertising and recommendation systems use boosting algorithms to better target ad-
verts and items to customers, depending on their preferences.
are an effective technique for many applications, such as customer segmentation, pre-
dictions, and fraud detection.
The random forests ensemble learning method builds a large number of decision
trees during the training phase and outputs the class that represents the mean of the
classes (classification) or mean prediction (regression) of the individual trees. Random
8 Meta-learning through ensemble approach 175
forests can be used for classification, regression, and other tasks. Due to their capacity
to restrict overfitting of a single decision tree and their efficiency in minimizing vari-
ance in the predictions, random forests have been demonstrated to be a highly accu-
rate and robust method for a range of situations [17].
Extra trees (sometimes referred to as extremely randomized trees), which ran-
domly choose threshold and feature values, and sparse random forests, which are in-
tended for high-dimensional datasets, are variations of the random forests algorithm.
Other variations include tailored random forests for specific tasks, such as random
forests for image classification and random forests for anomaly detection.
learners, ensemble methods can provide better performance and a better generaliza-
tion accuracy than a single model. The use of ensemble methods is applicable to both
classification and regression issues. Ensemble approaches frequently employ bagging,
boosting, stacking, and other strategies.
Machine learning employs the ensemble techniques of stacking and blending. Stack-
ing is an ensemble method that trains numerous base models on various subsets of a
single dataset. A single forecast for the dataset is then created by combining the pre-
dictions of each of these foundation models. Blending is an ensemble method that
combines estimates from various base models and trains a separate model to generate
the final estimate [19]. The weights used to combine the predictions from the basis
models are determined by how well the base models performed. Both stacking and
blending techniques can improve the accuracy of predictions as they combine the
power of multiple different models.
Ensemble pruning and model selection methods are techniques for selecting the best
possible model from among many different choices. The goal of pruning and model
selection is to either improve a model’s accuracy or to reduce the complexity of the
model, ensuring it is computationally tractable and efficient. Pruning involves remov-
ing irrelevant or noisy parameters from a model [20]. Model selection methods at-
tempt to select the best set of hyperparameters for a specific problem.
Ensembles methods involve combining multiple models that are connected to-
gether by various techniques. These techniques can include methods such as stacking,
averaging, sandboxing, boosting, and bagging. Ensemble pruning involves selecting
the subset of models that will provide the best performance. Model selection techni-
8 Meta-learning through ensemble approach 177
ques are used to select the most accurate model, given an input dataset. This can be
achieved by assessing accuracy scores, feature importance, and other metrics on the
trained model.
Ensemble learning is a powerful method for meta-learning tasks. Its goal is to com-
bine multiple models to create a consistent prediction. It works by aggregating multi-
ple weak learners to form a more powerful model. The general objective of ensemble
learning in meta-learning tasks is to combine multiple models into a single, stronger
model. The idea is to leverage the strengths of the individual models and to increase
the accuracy, while minimizing the risk of overfitting. This can be done by incorporat-
ing different techniques such as boosting, bagging or stacking, or combining different
types of algorithms, such as decision trees or neural networks [22]. Ensemble learning
can be used in meta-learning tasks to improve the accuracy of prediction. It can be
used to build comprehensive models with better performance, which can be accessed
more quickly. This makes it especially useful for time-critical tasks or tasks with high
volumes of data. Moreover, its robustness and ability to handle complex data sets
make it a great tool for meta-learning tasks.
178 S. Ramasamy et al.
Without the requirement for a second validation set or for dividing the initial dataset
into training and test sets, the supervised learning technique of cross-validation ena-
bles model evaluation and selection. Cross-validation is a technique that involves ran-
domly dividing the data into training and test sets; using the training set to build and
modify a model, and then using the test set to assess the model’s performance [24].
Cross-validation can be used to identify the best parameters and hyperparameters for
a given model, or to compare different algorithms.
Ensemble-based model evaluation is a technique for model evaluation in which
multiple models are trained and tested with different subsets of data. Each model is
then compared to a baseline model, and the performance of the ensemble is deter-
mined by its accuracy and precision on the held-out dataset. Ensembling techniques
can be used to combine prediction results from a variety of models, such as machine
learning algorithms, to improve predictive performance.
8 Meta-learning through ensemble approach 179
The underlying principles of a relational database and a NoSQL database are quite
different. A relational database is built upon the principles of relational-algebra,
which is a mathematical model for representing and manipulating data in a relational
way. This is why relational databases are often referred to as SQL, short for struc-
180 S. Ramasamy et al.
tured query language. Relational databases are organized into tables with rows and
columns for optimal data storage and retrieval speeds.
NoSQL databases are not built upon relational-algebra as their underlying princi-
ples. Instead, they are built upon the concepts of key-value stores, document data-
bases, and graph databases [25]. Compared to conventional relational databases, these
NoSQL databases offer a more flexible way to store and retrieve data. This adaptabil-
ity may result in quicker data storage and retrieval times.
The performance characteristics of both types of databases also vary significantly.
Relational databases are generally more robust when it comes to handling large
amounts of data. They are better suited for storing complex data structures and are
more resilient when it comes to scaling. On the other hand, NoSQL databases are more
suitable for handling unstructured data and are better at retrieving data quickly.
A potent machine learning approach called Gradient boosting machine (GBM) is used
to create predictive models from an ensemble of weak learners. It is a supervised
learning algorithm that hones on a weak learner, and then combines them in an en-
semble to create a stronger, more accurate model. GBM works by taking the errors
from the previously produced model as inputs in the new model. This in turn reduces
error and improves the overall model [27].
GBM comes in a variety of forms, including XGBoost, LightGBM, and CatBoost. A
distributed gradient-boosting toolkit called XGBoost uses parallelization to speed up
model training times. A novel gradient-boosting system called LightGBM makes use of
a tree-based learning technique to speed both training and inference. It also includes
a variety of cost functions, regularization, and shrinkage that further assist in reduc-
ing the model run time. Finally, CatBoost is a version of GBM that is specifically de-
signed to work with categorical data, allowing information to be uniquely encoded in
order to learn more effectively from a dataset.
issues, particularly those where a single model lacks adequate accuracy or predictive
ability, stacked ensembles are employed to increase prediction accuracy.
Ensemble pruning is a method for reducing the size of a predictive model by remov-
ing unnecessary input variables and increasing its generalization performance. By
eliminating unnecessary attributes or lowering the amount of attributes used in the
model, this method can be used to simplify prediction models. The ensemble model
can be more effective and accurate by choosing the most pertinent attributes. The
pruning process involves identifying the least important features and removing them
from the model. It can also be used to improve model interpretability by removing
attributes that do not have a significant contribution to the prediction [29]. Thresholds
can be used to define which attributes should be removed – such as removing attrib-
utes if the model accuracy is not decreased by a certain percentage. By using an en-
semble of models, the pruning process can be performed in parallel; multiple pruned
models can be combined to form a better-performing model than each individual
model alone.
Future directions:
– Implementing machine learning algorithms for predictive analytics
– Increasing the speed and scalability of processing large volumes of data
– Use of natural language processing technologies for extracting insights from data
– Creating an effective data governance system to manage the data
– Improving data visualization tools to increase user engagement
by using ensemble learning, but its effectiveness depends on being aware of the difficul-
ties and dangers the method presents.
In the future, meta-learning research will focus on improving the learning speed and
efficiency of meta-learning models while retaining the power of meta-knowledge. Re-
searchers are also trying to understand the role of transferable meta-knowledge by
combining it with domain knowledge, and this could lead to a better ability to handle
new tasks. Another topic of interest is the use of meta-learning for multi-task learning,
where the same meta-knowledge is used to solve different tasks. Finally, the ability to
use meta-learning methods in other areas of artificial intelligence is gaining traction,
such as in natural language processing, knowledge representation, and robotics.
tionally, further research could focus on boosting the predictive accuracy of ensemble
models by including more data sources, incorporating additional sources of domain
knowledge, or utilizing improved features.
References
[1] Olowookere, T. A., & Adewale, O. S. (2020). A framework for detecting credit card fraud with cost-
sensitive meta-learning ensemble approach. Scientific African, 8, e00464.
[2] de Paiva, B. B. M., Pereira, P. D., de Andrade, C. M. V., Gomes, V. M. R., Souza-Silva, M. V. R., Martins,
K. P. M. P., . . . & Marcolino, M. S. (2023). Potential and limitations of machine meta-learning
(ensemble) methods for predicting COVID-19 mortality in a large inhospital Brazilian dataset.
Scientific Reports, 13(1), 3463.
[3] Ma, Z., Guo, S., Xu, G., & Aziz, S. (2020). Meta learning-based hybrid ensemble approach for short-
term wind speed forecasting. IEEE Access, 8, 172859–172868.
[4] Yadav, S. P., Zaidi, S., Mishra, A., & Yadav, V. (2022). Survey on machine learning in speech emotion
recognition and vision systems using a recurrent neural network (RNN). Archives of Computational
Methods in Engineering, 29(3), 1753–1770.
[5] Guo, N., Di, K., Liu, H., Wang, Y., & Qiao, J. (2021). A metric-based meta-learning approach combined
attention mechanism and ensemble learning for few-shot learning. Displays, 70, 102065.
[6] Ahmed, K., Nadeem, M. I., Li, D., Zheng, Z., Al-Kahtani, N., Alkahtani, H. K., . . . & Mamyrbayev, O.
(2023). Contextually Enriched Meta-Learning Ensemble Model for Urdu Sentiment Analysis. Symmetry,
15(3), 645.
[7] Logeshwaran, J., & Karthick, S. (2022, April). A smart design of a multi-dimensional antenna to
enhance the maximum signal clutch to the allowable standards in 5G communication networks.
ICTACT Journal on Microelectronics, 8(1), 1269–1274.
[8] Zhang, J., Song, J., Gao, L., Liu, Y., & Shen, H. T. (2022). Progressive meta-learning with curriculum.
IEEE Transactions on Circuits and Systems for Video Technology, 32(9), 5916–5930.
[9] Yadav, S. P., & Yadav, S. (2020). Image fusion using hybrid methods in multimodality medical
images. Medical and Biological Engineering and Computing, 58, 669–687.
[10] Gastinger, J., Nicolas, S., Stepić, D., Schmidt, M., & Schülke, A. (2021, July). A study on ensemble
learning for time series forecasting and the need for meta-learning. In 2021 International Joint
Conference on Neural Networks (IJCNN), (pp. 1–8). IEEE.
[11] Gu, K., Zhang, Y., & Qiao, J. (2020). Ensemble meta-learning for few-shot soot density recognition.
IEEE Transactions on Industrial Informatics, 17(3), 2261–2270.
[12] Rajalakshmi, M., Saravanan, V., Arunprasad, V., Romero, C. T., Khalaf, O. I., & Karthik, C. (2022).
Machine learning for modeling and control of industrial clarifier process. Intelligent Automation
and Soft Computing, 32(1), 339–359.
[13] Raja, S., Logeshwaran, J., Venkatasubramanian, S., Jayalakshmi, M., Rajeswari, N., Olaiya, N. G., &
Mammo, W. D. (2022). OCHSA: Designing energy-efficient lifetime-aware leisure degree adaptive
routing protocol with optimal cluster head selection for 5G communication network disaster
management. Science Programme, 2022, 1–11
[14] Yang, A., Lu, C., Li, J., Huang, X., Ji, T., Li, X., & Sheng, Y. (2023). Application of meta-learning in
cyberspace security: A survey. Digital Communications and Networks, 9(1), 67–78.
[15] Anderson, R., Koh, Y. S., Dobbie, G., & Bifet, A. (2019). Recurring concept meta-learning for evolving
data streams. Expert Systems with Applications, 138, 112832.
8 Meta-learning through ensemble approach 187
[16] Cawood, P., & Van Zyl, T. (2022). Evaluating state-of-the-art, forecasting ensembles and meta-
learning strategies for model fusion. Forecasting, 4(3), 732–751.
[17] Shen, X., Dai, Q., Zhu, G., & Zhang, J. (2023). Dynamic ensemble pruning algorithms fusing meta-
learning with heuristic parameter optimization for time series prediction. Expert Systems with
Applications, 225, 120148.
[18] Hussain, S., & Ibrahim, N. (2022, June). Empirical investigation of role of meta-learning approaches
for the improvement of software development process via software fault prediction. In Proceedings
of the International Conference on Evaluation and Assessment in Software Engineering 2022,
(pp.413–420).
[19] Zhang, J., Song, J., Yao, Y., & Gao, L. (2021, October). Curriculum-based meta-learning. In
Proceedings of the 29th ACM International Conference on Multimedia, (pp.1838–1846).
[20] Saravanan, V., & Neeraja, A. (2013, January). Security issues in computer networks and
stegnography. In 2013 7th International Conference on Intelligent Systems and Control (ISCO),
(pp. 363–366). IEEE.
[21] Logeshwaran, J., Ramkumar, M., Kiruthiga, T., & Sharan Pravin, R. (2022, February). SVPA – the
segmentation based visual processing algorithm (SVPA) for illustration enhancements in digital
video processing (DVP). ICTACT Journal on Image and Video Processing, 12(3), 2669–2673.
[22] Kousik, N., Natarajan, Y., Raja, R. A., Kallam, S., Patan, R., & Gandomi, A. H. (2021). Improved salient
object detection using hybrid Convolution Recurrent Neural Network. Expert Systems with
Applications, 166, 114064.
[23] Yadav, S. P., Mahato, D. P., & Linh, N. T. D. (Eds.). (2020). Distributed artificial intelligence: A modern
approach. Boca Raton: CRC Press.
[24] Ma, S., & Fildes, R. (2021). Retail sales forecasting with meta-learning. European Journal of
Operational Research, 288(1), 111–128.
[25] Irfan, D., Tang, X., Narayan, V., Mall, P. K., Srivastava, S., & Saravanan, V. (2022). Prediction of Quality
Food Sale in Mart Using the AI-Based TOR Method. Journal of Food Quality, 2022, 1–13.
[26] Natarajan, Y., Kannan, S., Selvaraj, C., & Mohanty, S. N. (2021). Forecasting energy generation in
large photovoltaic plants using radial belief neural network. Sustainable Computing, 31, 100578.
[27] Yadav, S. P., Agrawal, K. K., Bhati, B. S., Al-Turjman, F., & Mostarda, L. (2022). Blockchain-based
cryptocurrency regulation: An overview. Computational Economics, 59(4), 1659–1675.
[28] Leka, H. L., Fengli, Z., Kenea, A. T., Hundera, N. W., Tohye, T. G., & Tegene, A. T. (2023). PSO-based
ensemble meta-learning approach for cloud virtual machine resource usage prediction. Symmetry,
15(3), 613.
[29] Khan, S., Saravanan, V., Gnanaprakasam, C. N., Jaya Lakshmi, T., Deb, N., Othman, N. A. (2022).
Privacy protection of healthcare data over social networks using machine learning algorithms.
Computational Intelligence and Neuroscience, 2022, 8, Article ID 9985933.
[30] Yuvaraj, N., Kousik, N. V., Jayasri, S., Daniel, A., & Rajakumar, P. (2019). A survey on various load
balancing algorithm to improve the task scheduling in cloud computing environment. Journal of
Advanced Research in Dynamical and Control Systems, 11(08), 2397–2406.
[31] Jain, A., Kumar, A., & Susan, S. (2022). Evaluating deep neural network ensembles by majority voting
cum meta-learning scheme. In Soft Computing and Signal Processing: Proceedings of 3rd ICSCSP
2020, (Vol. 2, pp. 29–37). Singapore: Springer.
[32] Logeshwaran, J., & Shanmugasundaram, R. N. (2019, December). Enhancements of resource
management for device to device (D2D) communication: A review. In 2019 Third International
conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud)(I-SMAC), (pp. 51–55). IEEE.
188 S. Ramasamy et al.
[33] Hannah, S., Deepa, A. J., Chooralil, V. S., BrillySangeetha, S., Yuvaraj, N., Arshath Raja, R., . . . &
Alene, A. (2022). Blockchain-based deep learning to process IoT data acquisition in cognitive data.
BioMed Research International, 2022, 1–7
[34] Dvornik, N., Schmid, C., & Mairal, J. (2019). Diversity with cooperation: Ensemble methods for few-
shot classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision
(pp.3723–3731).
[35] Zang, H., Cheng, L., Ding, T., Cheung, K. W., Wei, Z., & Sun, G. (2020). Day-ahead photovoltaic power
forecasting approach based on deep convolutional neural networks and meta learning.
International Journal of Electrical Power and Energy Systems, 118, 105790.
Abhisek Sethy, Nazeer Shaik, Pavan Kumar Yadavalli, S. P. Anandaraj
9 AI: issues, concerns, and ethical
considerations
Abstract: Artificial intelligence (AI) ethics are the ethical considerations arising from
the design and use of AI systems. Its focus is on understanding the ethical implications
of developing, deploying, and using AI systems, and on developing strategies to ensure
that AI systems are deployed responsibly and safely. AI ethics deals with issues such
as fairness, data privacy, and security. It also covers topics such as ethical algorithms,
ethical decision-making, trustworthiness, and transparency in AI.
9.1 Introduction
Artificial intelligence (AI) ethics is an area of philosophy that deals with the moral
implications of AI and how it affects society. It covers a broad range of topics includ-
ing privacy, autonomy, bias, fairness, ownership, responsibility, trust and transpar-
ency. It is a critical field to consider because AI is increasingly being used in many
different fields – from healthcare to finance – and its effects are far-reaching and
have a potential for significant impact. Many ethical considerations should be taken
into account in the development and deployment of AI technologies, such as ensuring
data privacy, preventing discrimination, and ensuring fairness [1]. Additionally, ethi-
cal principles should be embedded in the design and implementation of AI systems
and processes in order to ensure compliance with relevant laws and regulations as
well as overcome the potential for unintended consequences.
https://doi.org/10.1515/9783111323749-009
190 Abhisek Sethy et al.
equally and decisions are not based on bias or prejudice. Transparency makes sure
that users have visibility into how AI systems are operated and accountable allows
users to hold AI systems accountable for their decisions [2]. In essence, ethical consid-
erations in AI development and deployment protect people, businesses, and the tech-
nology itself from potential misuse or manipulation. As such, it is essential that
ethical considerations be taken into all aspects of AI development and deployment to
ensure the accurate and ethical use of AI technology.
attention to fairness metrics during the development process and formalize best practi-
ces for data collection and model testing.
Biased AI systems can have a serious effect on fairness and equity. AI systems can be
biased due to the data used to train them, the algorithms used to run them, and the
decision-making process. Biased AI systems can result in decision-making that is un-
fair, unjust, or exclusionary. For example, AI systems used for hiring decisions may
be more likely to select certain candidates based on gender or race. This can lead to a
lack of diversity in the workforce and reduce opportunities for those not included in
the algorithm [4]. It can also lead to decisions that are not in the best interests of the
individuals affected, such as through automated admissions systems. AI bias can have
far-reaching consequences for fairness and equity, both at an individual and societal
level. It can lead to a lack of equal opportunity, in terms of access to education, jobs,
or other opportunities. Additionally, AI bias can result in economic and social dispar-
ity, as advantages and disadvantages can be magnified across different groups. AI
bias can also lead to a reduced level of trust between people and organizations, as
people may feel their rights, security, or privacy are not being respected. In sum, AI
bias has implications for fairness and equity that are far-reaching and can lead to se-
rious consequences.
Fairness metrics and approaches in AI development are used to improve the fairness
of datasets and algorithms. They measure the performance of an algorithm on pro-
tected, sensitive characteristics such as gender or race [5]. By using fairness metrics,
AI developers can assess whether an AI model is producing different results for simi-
larly situated individuals and groups and is not treating some of them differently.
This helps ensure that the outcomes of AI models are fair and unbiased. Some of the
most popular fairness metrics and approaches include parity metrics, error rate par-
ity, power parity, goodness of fit, predictive value parity, and statistical parity.
These techniques are essential for developing and deploying explainable AI systems.
They can enable AI systems to provide more transparent and understandable results.
This would help in increasing user trust and well as fostering better acceptance and
usage of AI systems.
ant with regulations and laws. Properly implementing data privacy and protection
can protect businesses from data breaches, reputation damage, fines, and other nega-
tive consequences [9].
useful insights, but much of this data may contain sensitive information that needs to
be kept secure [11]. Privacy-preserving AI technology helps protect personal data while
also allowing AI algorithms to generate useful insights. Some common techniques used
include differential privacy, homomorphic encryption, federated learning, and secure
multiparty computation.
Legal and ethical frameworks for AI responsibility are based on laws and codes of
ethical conduct that govern the use of AI technology. These frameworks indicate how
AI systems should be developed, used, and deployed, in order to respect and uphold
198 Abhisek Sethy et al.
ethical standards and comply with laws and regulations related to AI. The frame-
works provide guidance on various aspects of AI technology and the responsible use
of AI, such as privacy, data security and safety, transparency, algorithmic decision-
making, accountability, and fairness. They may also address legal questions about lia-
bility and responsibility, as well as how scenarios involving AI technology and people
should be addressed. Laws and frameworks governing AI may also extent to AI-
enabled products, including those related to autonomous vehicles, healthcare technol-
ogies, facial recognition, and other services [19]. At the heart of any development,
deployment, or usage of AI technology is the importance of ensuring that user safety,
privacy, and security are primary considerations. AI systems must also be transparent
about their purpose, design, and function, while still safeguarding personal privacy
and guaranteeing anonymity when appropriate. Finally, AI systems must be held ac-
countable and evaluated for fairness and accuracy in automated decision-making pro-
cesses. The legal and ethical frameworks for AI responsibility seek to both protect
individuals and create an environment of fairness, trust, and responsibility when it
comes to how AI systems are developed, used, and deployed.
cess. Proper security measures can help protect data from unauthorized use, while
also protecting the system from malicious actors and attacks[22].
Ensuring robustness and resilience of AI systems is one of the most important chal-
lenges facing the AI community. Robustness is the ability of an AI system to perform
consistently across a variety of operating conditions. Resilience is the ability of an AI
system to be able to adapt and recover after being subjected to sudden changes in its
environment. To ensure robustness and resilience of AI systems, several approaches
can be followed. One approach involves testing AI systems during development to en-
sure they are able to perform in a variety of operating conditions. This helps identify
areas of improvement that can be worked on prior to deployment. Additionally, re-
searchers can use techniques such as reinforcement learning and deep learning to de-
velop AI models that are robust and resilient to changes in their environment. Such
models can be used to ensure that AI systems are able to adapt to sudden changes
and can continue to provide accurate outputs after being subjected to such changes
[23]. Furthermore, organizations can use safety protocols and protocols for monitor-
ing AI systems to ensure that they are operating within acceptable parameters. Addi-
tionally, organizations can use techniques such as data partitioning to help AI systems
identify anomalies more effectively, allowing them to adapt to changes in their envi-
ronment more quickly. Finally, AI systems can be designed to take into account the
fallibility of humans involved in their operation to ensure that AI systems are able to
respond to sudden changes.
Safeguarding AI systems against malicious intent involves securing the data, assump-
tions, algorithms, and infrastructure components used to create and deploy AI sys-
tems. Security teams should use a variety of techniques including access control,
identifying critical elements, and monitoring system operations to secure AI systems.
Furthermore, organizations should ensure that data used to train, test, and deploy AI
systems should be secure and that the data contains no malicious content [26]. This
includes rigorously testing for malicious code and protecting sensitive data. They
should also have robust procedures to identify malicious actors and thwart threats.
Finally, they should ensure that the infrastructure used to deploy AI systems is secure.
This includes securing network infrastructure and enforcing authentication and au-
thorization protocols. By taking these steps, organizations can ensure that AI systems
are resilient against malicious intent.
has the potential to improve productivity, reduce costs, and open up new business
opportunities. It can streamline processes, reduce labor costs, and generate more effi-
cient operations. It can also automate data entry, improve accuracy, and cut down on
paperwork. On the customer service side, AI technology can help create better cus-
tomer experiences with personalized recommendations and automated problem-
solving tools. In addition to the positive effects of AI technology, there are also poten-
tial risks. AI technology can increase unemployment, disrupt the labor market, and
raise ethical and legal concerns [27]. There are also worries that AI technologies could
lead to a loss of privacy as data is collected and used in unknown ways. There are
social changes to consider. AI technology can change the way people interact, leading
to a rise in virtual socializing. It could also have a wider impact on society by increas-
ing inequality, disrupting human interactions, and introducing bias into decision-
making processes. The social and economic impacts of AI technology are significant
and varied. Ultimately, it is important for society to explore these impacts and ensure
that the benefits outweigh any potential harms.
Evaluating the social and economic implications of AI involves looking at how AI im-
pacts individuals, organizations, and society as a whole. This means considering how AI
might affect the labor market, income inequality, privacy, security, and access to serv-
ices, as well as the ethical boundaries that should govern its use [28]. AI can lead to
profound shifts in the structures of the global economy and changes in how society
functions, enabling new possibilities in business and public services, but also leading to
new threats and vulnerabilities. Understanding the implications of AI on society and
the economy is essential for designing policies, regulations, and incentives that promote
beneficial applications while preventing or limiting the risks associated with AI.
in productivity as employers can reduce the amount of time it takes to complete tasks
and increase return on investments. This increases the cost effectiveness of the busi-
ness model and can lead to greater profits overall. It can also lead to the development
of new products and services that would not have been possible to develop before.
Automation is thus seen as an important component in business strategy. The automa-
tion can lead to increased job satisfaction as employees no longer need to complete
tedious and repetitive tasks, freeing up more time for creative tasks and problem solv-
ing [29]. This can lead to increased morale in the workplace and enhance productivity.
The automation has had a positive impact on businesses, from reducing the need for
manual labor to increasing productivity and job satisfaction. However, it has also had
a negative impact on some workers, reducing the number of manual labor jobs avail-
able and impacting those who may have been employed in those jobs.
AI and inequality are two closely related concepts. AI can help address disparities and
access for people across different segments of society. AI can help in mitigating dis-
parities and granting access to marginalized populations – for example, by providing
more personalized education. AI can also help in providing access to better healthcare
services and financial services for underserved populations. Additionally, AI can be
used to identify trends in a wide variety of datasets to better understand how inequal-
ity exists within communities. AI can also be used to develop policies and programs to
reduce disparities and increase access for vulnerable populations. Finally, AI can be
used to monitor the effectiveness of such policies and ensure that they are adequately
addressing the needs of all groups in society.
Incorporating ethical practices throughout the AI life cycle is a process of aligning eth-
ical principles with the development and deployment of AI. This involves identifying
ethical hazards throughout the design, training, deployment, and use of AI systems, as
well as actively promoting user safety and privacy. AI developers should strive to be
mindful of potential ethical fallouts from their designs, which may have adverse ef-
fects on individuals, communities, and society as a whole. In order to mitigate such
issues, developers can consider incorporating the following best practices into their
AI life cycle:
– Examine how the proposed AI system will affect users. AI researchers and devel-
opers should consider the potential risks posed by their AI system, such as the
potential for overfitting data that affects a user’s privacy or security.
– Consider ethical implications prior to deployment. AI developers should strive to
identify potential ethical issues and limit adverse effects before releasing AI into
the wild. This includes taking into account the potential unintended consequences
of an AI system, such as its prejudice against certain minority groups or its poten-
tially biased outcomes.
– Establish oversight and review processes. AI development teams should design an
oversight and review system in order to address ethical issues that may arise.
This can include developing an audit mechanism to check data inputs as well as
implementing measures for cross-checking outputs and making sure research
outcomes are consistent with indicated ethical principles.
9 AI: issues, concerns, and ethical considerations 205
and mechanisms are critical for addressing the ethical, legal, and societal implications
of AI technologies. They also provide guidance for building trust in AI systems, ensure
the responsible use of data, and promote transparency in decision-making. Ethical AI
governance frameworks and mechanisms aim to ensure that AI is used for the good
of humanity by balancing the benefits and potential risks associated with the technol-
ogy [35]. Examples of such frameworks and mechanisms include ethical AI codes of
conduct, data protection standards, and algorithmic fairness initiatives.
ensuring that data used in the AI system development is fair, unbiased, and accurate
and that the AI system is effectively monitored and regulated for its accuracy and any
adverse effects that it might produce. Ultimately, transparency and accountability in AI
governance ensures that AI systems are responsible and beneficial for everyone.
are being made [37]. AI systems and solutions must be designed with privacy and se-
curity in mind. As more sensitive data is being collected and shared with AI systems,
there is increased need for systems with strong security measures that protect the
data and personal information of individuals. This includes the development of secure
infrastructure, authentication measures, encryption, and other tools and techniques
that ensure the security of data.
Emerging ethical challenges in AI are those questions and dilemmas that arise as AI
and ML are increasingly used in our society and economy. These ethical questions
have become increasingly important as AI and ML algorithms are used to make deci-
sions and take action in areas ranging from medical care to financial services. The
questions include addressing potential risks around privacy, data security, accuracy,
accountability, and algorithmic bias, as well as the broader implications of applying
AI for social and economic purposes. It also necessitates a better understanding of the
legal, moral, and ethical implications of the technology. AI must be subject to stand-
ards that ensure outcomes are produced in a responsible and trustworthy manner,
taking into account the potential risks and benefits when a computer is deciding on
behalf of a person.
References
[1] Munoko, I., Brown-Liburd, H. L., & Vasarhelyi, M. (2020). The ethical implications of using artificial
intelligence in auditing. Journal of Business Ethics, 167, 209–234.
[2] Fiske, A., Henningsen, P., & Buyx, A. (2019). Your robot therapist will see you now: Ethical
implications of embodied artificial intelligence in psychiatry, psychology, and psychotherapy. Journal
of Medical Internet Research, 21(5), e13216.
[3] Keskinbora, K. H. (2019). Medical ethics considerations on artificial intelligence. Journal of Clinical
Neuroscience, 64, 277–282.
[4] Varlamov, O. O., Chuvikov, D. A., Adamova, L. E., Petrov, M. A., Zabolotskaya, I. K., & Zhilina,
T. N. (2019). Logical, philosophical and ethical aspects of AI in medicine. International Journal of
Machine Learning and Computing, 9(6), 868.
[5] Giuffrida, I. (2019). Liability for AI decision-making: Some legal and ethical considerations. Fordham
Law Review, 88, 439.
[6] Schönberger, D. (2019). Artificial intelligence in healthcare: A critical analysis of the legal and ethical
implications. International Journal of Law and Information Technology, 27(2), 171–203.
[7] Holmes, W., Porayska-Pomsta, K., Holstein, K., Sutherland, E., Baker, T., Shum, S. B., . . . &
Koedinger, K. R. (2021). Ethics of AI in education: Towards a community-wide framework.
International Journal of Artificial Intelligence in Education, 2021, 1–23.
[8] Safdar, N. M., Banja, J. D., & Meltzer, C. C. (2020). Ethical considerations in artificial intelligence.
European Journal of Radiology, 122, 108768.
[9] Hancock, J. T., Naaman, M., & Levy, K. (2020). AI-mediated communication: Definition, research
agenda, and ethical considerations. Journal of Computer-Mediated Communication, 25(1), 89–100.
[10] Ouchchy, L., Coin, A., & Dubljević, V. (2020). AI in the headlines: The portrayal of the ethical issues of
artificial intelligence in the media. AI and Society, 35, 927–936.
[11] Borenstein, J., & Howard, A. (2021). Emerging challenges in AI and the need for AI ethics education.
AI and Ethics, 1, 61–65.
[12] Liebrenz, M., Schleifer, R., Buadze, A., Bhugra, D., & Smith, A. (2023). Generating scholarly content
with ChatGPT: Ethical challenges for medical publishing. The Lancet Digital Health, 5(3), e105–e106.
210 Abhisek Sethy et al.
[13] Garrett, N., Beard, N., & Fiesler, C. (2020, February). More than “If Time Allows” the role of ethics in
AI education. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (pp. 272–278).
[14] Murphy, K., Di Ruggiero, E., Upshur, R., Willison, D. J., Malhotra, N., Cai, J. C., . . . & Gibson, J. (2021).
Artificial intelligence for good health: A scoping review of the ethics literature. BMC Medical Ethics,
22(1), 1–17.
[15] Floridi, L., Cowls, J., King, T. C., & Taddeo, M. (2020). How to design AI for social good: Seven
essential factors. Science and Engineering Ethics, 26, 1771–1796.
[16] Yadav, S. P., & Yadav, S. (2019). Mathematical implementation of fusion of medical images in
continuous wavelet domain. Journal of Advanced Research in Dynamical and Control System, 10(10),
45–54.
[17] Logeshwaran, J., Kiruthiga, T., Kannadasan, R., Vijayaraja, L., Alqahtani, A., Alqahtani, N., & Alsulami,
A. A. (2023). Smart load-based resource optimization model to enhance the performance of device-
to-device communication in 5G-WPAN. Electronics, 12(8), 1821.
[18] Rani, P., Verma, S., Yadav, S. P., Rai, B. K., Naruka, M. S., & Kumar, D. (2022). Simulation of the
lightweight blockchain technique based on privacy and security for healthcare data for the cloud
system. International Journal of E-Health and Medical Communications (IJEHMC), 13(4), 1–15.
[19] Yadav, S. P. (2022). Blockchain security. Blockchain Security in Cloud Computing, 1–18.
[20] Pujahari, R. M., Yadav, S. P., & Khan, R. (2022). Intelligent farming system through weather forecast
support and crop production. In Application of Machine Learning in Agriculture (pp. 113–130). India:
Academic Press.
[21] Morris, M. R. (2020). AI and accessibility. Communications of the ACM, 63(6), 35–37.
[22] Hagendorff, T. (2020). The ethics of AI ethics: An evaluation of guidelines. Minds and Machines,
30(1), 99–120.
[23] Belk, R. (2021). Ethical issues in service robotics and artificial intelligence. The Service Industries
Journal, 41(13–14), 860–876.
[24] Char, D. S., Abràmoff, M. D., & Feudtner, C. (2020). Identifying ethical considerations for machine
learning healthcare applications. The American Journal of Bioethics, 20(11), 7–17.
[25] Morley, J., Machado, C. C., Burr, C., Cowls, J., Joshi, I., Taddeo, M., & Floridi, L. (2020). The ethics of AI
in health care: A mapping review. Social Science and Medicine, 260, 113172.
[26] Subramanian, B., Saravanan, V., Nayak, R. K., Gunasekaran, T., & Hariprasath, S. (2019). Diabetic
retinopathy-feature extraction and classification using adaptive super pixel algorithm. International
Journal of Engineering and Advanced Technology, 9, 618–627.
[27] Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J., . . . &
Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey
wolf optimization in wireless sensor networks. Security and Communication Networks, 2022, 1–10
[28] Tripathy, R., Nayak, R. K., Saravanan, V., Mishra, D., Parasa, G., Das, K., & Das, P. (2021). Spectral
clustering based fuzzy C-means algorithm for prediction of membrane cholesterol from ATP-binding
cassette transporters. In Intelligent and Cloud Computing: Proceedings of ICICC 2019 (Vol. 2,
pp. 439–448). Singapore: Springer.
[29] Esmaeilzadeh, P. (2020). Use of AI-based tools for healthcare purposes: A survey study from
consumers’ perspectives. BMC Medical Informatics and Decision Making, 20(1), 1–19.
[30] Reddy, S., Allan, S., Coghlan, S., & Cooper, P. (2020). A governance model for the application of AI in
health care. Journal of the American Medical Informatics Association, 27(3), 491–497.
[31] Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model
for efficient resource allocation in 5G broadband communication networks. ICTACT Journal On Soft
Computing, 13(3), 2986–2994.
[32] Mittelstadt, B. (2019). Principles alone cannot guarantee ethical AI. Nature Machine Intelligence,
1(11), 501–507.
9 AI: issues, concerns, and ethical considerations 211
[33] Morley, J., Floridi, L., Kinsey, L., & Elhalal, A. (2020). From what to how: An initial review of publicly
available AI ethics tools, methods and research to translate principles into practices. Science and
Engineering Ethics, 26(4), 2141–2168.
[34] Ponnusamy, M., Bedi, P., Suresh, T., Alagarsamy, A., Manikandan, R., & Yuvaraj, N. (2022). Design
and analysis of text document clustering using salp swarm algorithm. The Journal of
Supercomputing, 78(14), 16197–16213.
[35] Kannan, S., Yuvaraj, N., Idrees, B. A., Arulprakash, P., Ranganathan, V., Udayakumar, E., & Dhinakar, P.
(2021). Analysis of convolutional recurrent neural network classifier for COVID-19 symptoms over
computerised tomography images. International Journal of Computer Applications in Technology, 66
(3–4), 427–432.
[36] Manikandan, V., & Yuvaraj, N. (2019, May). Advanced expert system using particle swarm
optimization based adaptive network based fuzzy inference system to diagnose the physical
constitution of human body. In Emerging Technologies in Computer Engineering: Microservices in
Big Data Analytics: Second International Conference, ICETCE 2019, Jaipur, India, February 1–2, 2019,
Revised Selected Papers (Vol. 985, p. 349). Springer.
[37] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). Energy‐efficient resource allocation
model for device‐to‐device communication in 5G wireless personal area networks. International
Journal of Communication Systems, 2023, 1–14.
[38] Schiff, D., Biddle, J., Borenstein, J., & Laas, K. (2020, February). What’s next for AI ethics, policy, and
governance? A global overview. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and
Society (pp. 153–158).
Abhisek Sethy, Nazeer Shaik, V. Saravanan, S. Karthika
10 The future with AI and AI in action
Abstract: The future of artificial intelligence (AI) is bright and filled with possibilities.
AI has the potential to revolutionize many areas, including medicine, finance, cus-
tomer service, and transportation. It can help create more efficient systems, enhance
decision-making, and ultimately improve people’s lives. With increased computing
power and faster networks, AI can become more sophisticated. Eventually, AI could
be used to automate mundane and labor-intensive tasks such as filing paperwork or
scheduling meetings. AI can also help identify new opportunities for businesses and
increase their competitiveness. For example, AI can be used for data mining to allow
organizations to make more informed decisions. The AI can be used to create more
realistic simulations and virtual environments to help with the development of prod-
ucts, software, and autonomous cars. In the future, AI might even be able to take on
more complex tasks such as visual recognition, language translation, and natural lan-
guage processing. Ultimately, AI will play an increasingly important role in how busi-
nesses operate and how people interact with each other. As AI technology evolves, so
too will its applications in the future.
AI is an area of computer science that has seen rapid advancements in recent years due
to the availability of data, cloud computing, and powerful algorithms. AI technologies
are being used to solve a variety of complex problems related to automation, natural
https://doi.org/10.1515/9783111323749-010
214 Abhisek Sethy et al.
10.2 AI in healthcare
AI in healthcare is the utilization of advanced AI techniques to provide innovative and
more efficient medical solutions. AI can help improve patient outcomes, enhance the
safety of medical treatments, reduce costs, and improve patient satisfaction. AI can also
help healthcare organizations improve diagnostic accuracy, patient flow, provider com-
munication, and operational efficiency of their organizations [4]. AI can be used to iden-
tify patterns in medical records, develop better prediction models to diagnose and treat
diseases, automate some administrative and care processes, and provide personalized
and preventive care. AI also allows for easier data sharing, increases the accuracy of
medical records, allows for more effective patient-provider communication, and can
provide streamlined medical care coordination.
AI is a rapidly growing field and has the potential to revolutionize the healthcare in-
dustry. AI can be used in healthcare for a variety of purposes, including diagnostics,
prognostics, personalized healthcare, and advanced medical treatments. AI can be
employed to detect and diagnose medical conditions, utilize data and predict out-
comes, monitor patients, provide personalized treatments, and even suggest the best
treatment options for certain diseases. AI can also be used to analyze vast amounts of
patient data to improve disease management and help determine the most suitable
patient-specific treatments [5]. Additionally, AI can be used to improve the perfor-
mance and accuracy of medical imaging techniques such as CT scans and MRIs. AI
can even be used to plan and deliver targeted drug delivery and treatment. AI can
also be used to build intelligent robots and machines to assist surgeons and other
medical personnel in performing medical procedures. AI has already been imple-
mented in some hospitals and clinics, and its uses in healthcare are growing fast. As
AI-equipped robots, machines and algorithms become increasingly advanced, they
can be employed to perform a greater range of tasks, from diagnostics to surgery.
AI-enabled medical imaging and diagnostics is the use of AI to analyze medical im-
ages, such as MRI scans and X-rays, for the purpose of diagnosing or detecting medical
problems. AI-enabled medical imaging and diagnostics can help reduce the time spent
manually analyzing images, improve accuracy and reproducibility of diagnoses, and
even highlight abnormalities in a patient’s medical images that may otherwise be
missed. AI-enabled medical imaging and diagnostics can also reduce errors associated
with misinterpretation of medical images [6]. Many large medical institutions have
216 Abhisek Sethy et al.
AI-assisted drug discovery and personalized medicine are two closely related concepts
in healthcare that are driven by advances in AI. AI-assisted drug discovery is an
emerging field that combines ML and AI to enable drug discovery and development
faster and more efficiently. This can be used to identify and develop new treatments
for a wide range of illnesses with fewer resources and in less time, leading to fewer
expensive failures and a greater likelihood of success. Personalized medicine is an ap-
proach to healthcare that uses an individual’s genetic and nongenetic traits to better
anticipate, diagnose, and treat various illnesses [7]. This approach can provide person-
alized treatments to fit an individual’s specific needs, which can improve patient out-
comes and lead to lower health costs. AI-assisted drug discovery can be combined
with personalized medicine to identify treatments that are tailored to each individual
patient, allowing for improved efficacy and better patient outcomes.
10.3 AI in finance
AI in finance has become a major factor in how companies, financial institutions, and
organizations manage their resources. AI provides a more efficient way to make deci-
sions and process data, allowing financial institutions to identify and seize market op-
portunities. AI can also be used to automate mundane and time-consuming tasks in
the financial industry, such as compliance and fraud prevention, capital management,
and portfolio management. Additionally, AI has been used to create algorithmic trad-
ing strategies, simulate risk scenarios, and develop optimized stock portfolios. AI can
also be used to automate customer service and provide predictive analytics for cus-
tomer segmentation and personalization. AI is quickly becoming an essential tool for
financial institutions to stay competitive and maximize their profits.
analyze market conditions and trends, and identify potential trading opportunities. AI
algorithms can also assist with optimizing allocations to help maximize profits and
reduce risk [10].
Regulatory considerations and challenges in AI Finance (AI FinReg) are the legal, reg-
ulatory, and ethical requirements and potential implications associated with the use
of AI in finance. AI-based financial systems create new risks associated with transpar-
ency, fairness, reliability, safety, and security. Financial regulators are beginning to
develop regulations to address these risks. For instance, the European Union’s General
Data Protection Regulation (GDPR) provides guidelines for data protection and pri-
vacy considerations when developing and deploying AI-based financial services. In
the United States, the Office of the Comptroller of the Currency (OCC) has developed
an anticipatory framework for handling regulatory issues associated with the use of
AI in financial services [11]. In addition, regulators are actively engaged in conversa-
tions and consultations with industry stakeholders to discuss the use of AI in financial
services. These conversations often reveal potential blind spots in the existing regula-
tory framework and suggest potential solutions. Despite these proactive measures, the
regulatory landscape is difficult to navigate because AI technology is still so new and
dynamic. Regulators will continue to assess financial services driven by AI, so the
exact regulations and enforcement of these regulations are continuously evolving. As
such, it is important for businesses to remain abreast of relevant regulations and stay
connected with industry leaders to ensure compliance.
10 The future with AI and AI in action 219
10.4 AI in transportation
AI is increasingly being used in the transportation sector to create smarter and more
efficient transportation networks. AI can be used to enhance safety, reduce traffic ac-
cidents, expand capacity, improve customer experience, and reduce costs. It can help
enable autonomous, self-driving vehicles; optimize route planning and maintenance,
and facilitate the integration of alternative energy sources. AI has also been used to
analyze data collected from vehicle sensors, detect traffic patterns, and predict future
conditions. By leveraging AI-based algorithms, transportation networks can be made
more reliable, efficient, and environmentally friendly.
Quality control and defect detection using AI is a process aimed at ensuring that man-
ufactured products or services meet the required specifications. AI-based solutions
can help identify quality and defect problems quickly and accurately, reducing the
time and cost associated with conventional methods. AI-based quality control tools in-
corporate object recognition algorithms that can autonomously process the visualized
output of digital inspection equipment and compare it against predefined criteria in
order to detect defects. AI-based systems can also learn over time from production
data and detect patterns in the occurrence of quality or defects across products and
processes, leading to improved process control and quality assurance.
222 Abhisek Sethy et al.
analytics, and ML can provide businesses with data-driven insights into customer
preferences and trends that help marketers develop more effective strategies. AI can
also be used to customize customer experiences based on specific needs and interests,
making them more personalized and engaging. Additionally, AI can automate market-
ing tasks such as segmentation, targeting, and personalization, saving businesses time
and resources. By leveraging the power of AI, businesses can drive customer engage-
ment and loyalty while creating more efficient and effective marketing strategies.
Sentiment analysis is a technique used in the field of NLP to identify the opinion or
attitude of a speaker. AI-driven sentiment analysis works by interpreting text or
audio data to detect underlying sentiment. In customer sentiment tracking, AI-driven
sentiment analysis can be used to quickly and accurately identify customer attitudes
toward a brand or product. AI’s ability to quickly process large amounts of data from
multiple sources allows for detailed analysis and reporting to be generated within mi-
nutes. For example, customer sentiment tracking can be used to analyze customer
feedback from social media posts, product and service reviews, surveys, and more to
gain an understanding of customer attitudes [21]. AI-driven sentiment analysis can
also detect sentiment in real time, allowing for more effective customer service and
issue resolution. By combining sentiment analysis with AI-driven algorithms and ML,
businesses are able to quickly identify customer pain points, complaints, and feed-
back. This data can then be used to inform product or service changes, customer ser-
vice strategies, and marketing campaigns. Ultimately, AI-driven sentiment analysis is
an invaluable tool for businesses aiming to provide better customer experiences.
224 Abhisek Sethy et al.
10.7 AI in education
AI in education is the use of AI technologies in the classroom. AI technologies can be
used to personalize learning experiences, improve student engagement, and automate
routine activities. AI can also be used to support teaching by providing real-time feed-
back to students on their progress and identifying areas of improvement. It can also
help instructors be more efficient and provide a better learning environment for stu-
dents. AI can also be used to create digital tutors and virtual classrooms that assist in
delivering personalized instruction. AI-driven technologies such as bots, virtual assis-
tants, and intelligent tutoring systems allow for more efficient learning and knowl-
edge acquisition of key educational concepts.
10 The future with AI and AI in action 225
Adaptive learning platforms are systems that use AI to determine when and how to de-
liver content to students. AI tutors use NLP and ML techniques to understand a student’s
current abilities, goals, and educational preferences. They then generate assignments, ac-
tivities, and resources that are tailored to the student’s individual needs. AI tutors can
deliver content in various forms, such as videos, interactive games, online worksheets,
and more. In addition, they provide feedback, take tests, and monitor student progress
and engagement. AI tutors are used to supplement or replace traditional instructors, al-
lowing for personalized learning experiences and real-time personalized instruction.
their crops. Farmers can adjust their irrigation, fertilization, and management strate-
gies accordingly, to ensure that their crops are not only healthy but also perform
optimally.
Pest detection and disease monitoring using AI is a process whereby AI is used to de-
tect, diagnose, and track pests and diseases in agriculture. This can be done in a vari-
ety of ways. AI can gather data about pests and diseases from multiple sources and
use the data to identify potential outbreaks and track changes in the environment. It
can also use algorithms to identify patterns in the data, helping to identify when ac-
tion needs to be taken. AI can also provide detailed maps of the risk factors associated
with a particular pest or disease and help develop strategies for prevention and
management.
AI applications in food processing and quality control involve the use of AI-powered
analytics, such as computer vision and ML to automate and optimize the processes of
food production and inspection. For example, AI can be used to detect quality anoma-
lies in food products and enable faster and more accurate quality decision-making. AI
can also be used to automate the production process and reduce labor costs. Addition-
ally, AI-driven analytics can be used to improve predictive analytics and supply chain
performance. AI can help reduce food waste by detecting anomalies in production op-
erations and making sure only the highest quality products reach consumers.
AI for urban planning and resource optimization involves using AI algorithms and
systems to improve the planning and management of urban areas and to optimize the
use of resources. The goal is to create more efficient and sustainable cities. This in-
cludes the use of ML algorithms and models to analyze population size, zoning laws,
transit patterns, economic development, and other factors to determine the best way
to plan and build a city. AI can also be used to streamline and optimize the allocation
of resources to areas with a higher demand. This can then help reduce costs and pro-
vide better quality of life to citizens. Additionally, AI can reduce and minimize the
amount of waste produced by cities by optimizing processes like energy consumption,
water management, and waste management.
ally, ITS systems can predict traffic route recommendations, route guidance, and trip
times. By helping optimize traffic flow, reduce emissions, and save time and money,
ITS is a key component in creating sustainable and dynamic smart cities.
AI-powered energy management and sustainability is a term used to refer to the use
of AI and ML algorithms to empower organizations to become more responsible and
use resources more efficiently. AI-powered energy management and sustainability al-
lows businesses to optimize their processes by analyzing a variety of data points re-
lated to energy consumption and corporate sustainability. The result is a system that
can identify and implement better practices that reduce energy costs and minimize
environmental impact [28].
AI-powered energy management and sustainability solutions provide companies
with tremendous control over usage and consumption of energy and resources. This
type of solution can be used to monitor and track energy usage, environmental distur-
bance, and irregularities in energy usage from areas without proper tracking. By
monitoring these metrics, businesses can identify potential problems and resolve
them in a timely fashion. Additionally, AI-powered energy management and sustain-
ability solutions can help organizations optimize energy usage and consumption pref-
erences to maximize energy efficiency and reduce costs.
AI-powered energy management and sustainability solutions can also provide
global oversight that aids in regulatory compliance, which allows businesses to be
proactive rather than reactive. Such a system can enable companies to implement
new strategies for reducing energy consumption while working to achieve corporate
sustainability objectives. AI-powered energy management and sustainability solutions
are becoming increasingly popular due to their comprehensive features and scalabil-
ity [29].
Privacy, security, and ethical concerns in AI smart cities are issues that need to be ad-
dressed as cities become increasingly reliant on technology to provide services such as
transportation and waste management. With the increased reliance on AI, ethical con-
cerns have arisen, such as potential impacts on human rights, privacy, data protection,
algorithmic accountability, and market competition. Ethical issues are particularly im-
portant, as AI systems can have significant effects on people’s lives [30].
The implications of AI and the technology that is used to create these cities must
be studied to assess whether or not they are adequately protecting individuals’ pri-
vacy and other human rights. A key concern is the accuracy of data used by AI, as
230 Abhisek Sethy et al.
inaccurate data can lead to biased decisions that could affect people’s lives [31]. Addi-
tionally, AI algorithms have the potential to strengthen existing inequality by amplify-
ing existing patterns of discrimination or favoring certain user interests. This bias in
decision-making could further marginalize certain groups if not considered early on.
In terms of security, AI smart cities are particularly vulnerable to attack due to the
reliance on digital technologies and the speed at which information is relayed [32]. As
a result, cities will need to develop extensive security systems to protect themselves
from malicious actors using cyberattacks to disrupt services, steal data, or create secu-
rity vulnerabilities. Ultimately, AI smart cities have the potential to provide significant
benefits for their citizens as long as they consider the privacy, security, and ethical
implications of their technology. It is vital that cities plan carefully in order to ensure
that their AI algorithms and systems are fair and accountable, while simultaneously
protecting the privacy of individuals and creating secure networks [33].
References
[1] Haenlein, M., & Kaplan, A. (2019). A brief history of artificial intelligence: On the past, present,
and future of artificial intelligence. California Management Review, 61(4), 5–14.
[2] Zhang, C., & Lu, Y. (2021). Study on artificial intelligence: The state of the art and future prospects.
Journal of Industrial Information Integration, 23, 100224.
[3] Schwalbe, N., & Wahl, B. (2020). Artificial intelligence and the future of global health. The Lancet,
395(10236): 1579–1586.
[4] Schiff, D. (2021). Out of the laboratory and into the classroom: The future of artificial intelligence in
education. AI and Society, 36(1): 331–348.
[5] Belfield, H. (2020, February). Activism by the AI community: Analysing recent achievements and
future prospects. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (pp. 15–21).
[6] Biswas, S., Carson, B., Chung, V., Singh, S., & Thomas, R. (2020). AI-Bank of the future: Can banks
meet the AI challenge. New York: McKinsey & Company.
[7] Cave, S. (2020, February). The problem with intelligence: Its value-laden history and the future of AI.
In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (pp. 29–35).
[8] Sur, J., Bose, S., Khan, F., Dewangan, D., Sawriya, E., & Roul, A. (2020). Knowledge, attitudes, and
perceptions regarding the future of artificial intelligence in oral radiology in India: A survey.
Imaging Science in Dentistry, 50(3): 193.
[9] Aggarwal, K., Mijwil, M. M., Al-Mistarehi, A. H., Alomari, S., Gök, M., Alaabdin, A. M. Z., &
Abdulrhman, S. H. (2022). Has the future started? The current growth of artificial intelligence,
machine learning, and deep learning. Iraqi Journal for Computer Science and Mathematics,
3(1): 115–123.
[10] Sangeetha, R., Logeshwaran, J., Rocher, J., & Lloret, J. (2023). An improved agro deep learning model
for detection of Panama Wilts disease in banana leaves. AgriEngineering, 5(2), 660–679.
[11] Singh, P. K., Chauhan, S. S., Sharma, A., Prakash, S., & Singh, Y. (2023). Prediction of higher heating
values based on imminent analysis by using regression analysis and artificial neural network for
bioenergy resources. In Proceedings of the Institution of Mechanical Engineers, Part E: Journal of
Process Mechanical Engineering. 09544089231175046.
10 The future with AI and AI in action 231
[12] Chauhan, S. S., Yadav, S. P., Awashthi, S., & Naruka, M. S. (2023). Sentimental analysis using cloud
dictionary and machine learning approach. In: Kaushal Kishor, Neetesh Saxena & Dilkeshwar
Pandey. Cloud-based Intelligent Informative Engineering for Society 5.0 (pp. 157–169). Boca Raton:
Chapman and Hall/CRC.
[13] Awasthi, S., Naruka, M. S., Yadav, S. P., & De Albuquerque, V. H. C. (Eds.). (2023). AI and IoT-based
Intelligent Health Care & Sanitation. United Arab Emirates: Bentham Science Publishers.
[14] Yadav, S. P., Zaidi, S., Nascimento, C. D. S., de Albuquerque, V. H. C., & Chauhan, S. S. (2023,
January). Analysis and design of automatically generating for GPS Based Moving Object Tracking
System. In 2023 International Conference on Artificial Intelligence and Smart Communication (AISC)
(pp. 1–5). IEEE.
[15] Saravanan, V., Sumathi, A., Shanthana, S., & Rizvana, M. (2013). Dual mode mpeg steganography
scheme for mobile and fixed devices. International Journal of Engineering Research and
Development, 6, 23–27.
[16] Cihon, P., Maas, M. M., & Kemp, L. (2020). Fragmentation and the future: Investigating architectures
for international AI governance. Global Policy, 11(5): 545–556.
[17] Stahl, B. C. (2021). Artificial intelligence for a better future: An ecosystem perspective on the ethics
of AI and emerging digital technologies (p. 124). New York: Springer Nature.
[18] Allam, Z., Dey, G., & Jones, D. S. (2020). Artificial intelligence (AI) provided early detection of the
coronavirus (COVID-19) in China and will influence future Urban health policy internationally. Ai,
1(2): 156–165.
[19] Saravanan, V., & Chandrasekar, C. (2012). Qos-continuous live media streaming in mobile environment
using vbr and edge network. International Journal of Computer Applications, 53(6), 1–3.
[20] Lakshminarayanan, R., & Yuvaraj, N. (2020). Design and analysis of an improved deep learning
algorithm on classification of intervertebral discs. International Journal of Advanced Science and
Technology, 29, 4019–4026.
[21] Gupta, S., Saravanan, V., Choudhury, A., Alqahtani, A., Abonazel, M. R., & Babu, K. S. (2022).
Supervised computer-aided diagnosis (CAD) methods for classifying Alzheimer’s disease-based
neurodegenerative disorders. Computational and Mathematical Methods in Medicine, 2022, 1–10
[22] Indhumathi, R., Amuthabala, K., Kiruthiga, G., Yuvaraj, N., & Pandey, A. (2023). Design of task
scheduling and fault tolerance mechanism based on GWO algorithm for attaining better QoS in
cloud system. Wireless Personal Communications, 128(4): 2811–2829.
[23] Zhang, K., & Aslan, A. B. (2021). AI technologies for education: Recent research & future directions.
Computers and Education: Artificial Intelligence, 2, 100025.
[24] Vaishya, R., Javaid, M., Khan, I. H., & Haleem, A. (2020). Artificial Intelligence (AI) applications for
COVID-19 pandemic. Diabetes and Metabolic Syndrome: Clinical Research and Reviews,
14(4): 337–339.
[25] Aruna, R. D., Surendran, S. D., Yuvaraj, N. D., & Debtera, B. (2022). An enhancement on
convolutional Artificial Intelligent based diagnosis for skin disease using nanotechnology sensors.
Computational Intelligence and Neuroscience, 2022, 1–7.
[26] Berente, N., Gu, B., Recker, J., & Santhanam, R. (2021). Managing artificial intelligence. MIS
Quarterly, 45(3), 1–18.
[27] Yuvaraj, N., Praghash, K., & Karthikeyan, T. (2022). Privacy preservation of the user data and
properly balancing between privacy and utility. International Journal of Business Intelligence and
Data Mining, 20(4): 394–411.
[28] Ramesh, G., Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). Prediction of energy production level
in large PV plants through AUTO-Encoder Based Neural-Network (AUTO-NN) with restricted
Boltzmann feature extraction. Future Internet, 15(2), 46.
232 Abhisek Sethy et al.
[29] Saravanan, V., Sreelatha, P., Atyam, N. R., Madiajagan, M., Saravanan, D., & Sultana, H. P. (2023).
Design of deep learning model for radio resource allocation in 5G for massive IoT device.
Sustainable Energy Technologies and Assessments, 56, 103054.
[30] Borges, A. F., Laurindo, F. J., Spínola, M. M., Gonçalves, R. F., & Mattos, C. A. (2021). The strategic use
of artificial intelligence in the digital era: Systematic literature review and future research directions.
International Journal of Information Management, 57, 102225.
[31] Lu, Y. (2019). Artificial intelligence: A survey on evolution, models, applications and future trends.
Journal of Management Analytics, 6(1): 1–29.
[32] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). L-RUBI: An efficient load-based
resource utilization algorithm for bi-partite scatternet in wireless personal area networks.
International Journal of Communication Systems, 36(4), 1–17.
[33] Fouad, H., Hassanein, A. S., Soliman, A. M., & Al-Feel, H. (2020). Analyzing patient health information
based on IoT sensor with AI for improving patient assistance in the future direction. Measurement,
159, 107757.
S. P. Sreekala, S. Revathy, S. Rajeshwari, B. Girimurugan
11 A survey of AI in industry: from
basic concepts to industrial and
business applications
Abstract: Artificial intelligence (AI) has the potential to revolutionize nearly every in-
dustry, and is already making a significant impact on many. AI in the enterprise has
proven to be a major influence on the way products and services are developed, deliv-
ered, and sold. AI has helped to automate mundane manual processes, as well as en-
able companies to provide personalized customer experiences. In the manufacturing
industry, AI has enabled robots to replace manual labor, resulting in increased output
and improved efficiency. In the retail sector, AI is being used for product recommen-
dations, increased customer personalization, and predictive analytics for inventory
management. AI is also being used in the healthcare industry to improve patient diag-
nosis, predict medical outcomes, and analyze images for potential anomalies. AI has
made many areas of industry more efficient and has opened up the possibility of
many more advancements.
https://doi.org/10.1515/9783111323749-011
234 S. P. Sreekala et al.
The use of AI in industrial and business settings is rapidly growing and becoming in-
creasingly important due to the potential it offers for improving efficiency, productiv-
ity, and cost savings. AI can match the demand for personalized services and tailored
products that are tailored to the specific needs of customers. AI also helps in making
more accurate analyses and forecasts, thereby helping to identify areas for improve-
ment and avoid potential risks. Moreover, AI can speed up decision-making, automate
operations, and provide guidance on areas that could be improved [3]. AI is used in a
number of industries, including finance and healthcare, and a wide range of business
functions such as marketing, human resources, and operations. AI is becoming ubiq-
uitous in businesses and industries as it has the potential to revolutionize the way
companies conduct their business.
AI principles and techniques refer to the specific ways that AI can be used to solve
problems and provide solutions. These principles and techniques can range from al-
gorithms, machine learning, natural language processing (NLP), robotics, and more
[5]. The principles of AI, such as problem solving, search, planning, knowledge repre-
sentation, and learning, are fundamental to the development of successful AI systems.
AI techniques, such as supervised and unsupervised learning, neural networks, Bayes-
ian networks, and RL, allow for more refined and sophisticated applications of these
principles. Along with the advancements of deep learning, more complex AI systems
can be developed to tackle complex problems and derive highly accurate predictive
models.
Machine learning algorithms are a set of algorithms used to identify patterns in data
and learn from such data. The main purpose is to enable computers to learn from
data and use these learnings to create predictive models and automate decisions.
These algorithms are broadly categorized into supervised learning, unsupervised
learning, and RL:
– Supervised learning algorithms are used to predict the output, given a certain
input. Examples of supervised algorithms include linear regression, decision
trees, logistic regression, and support vector machines.
– Unsupervised learning algorithms is used to identify patterns in data without la-
bels. Examples of unsupervised algorithms include clustering, dimensionality re-
duction, and anomaly detection.
– RL algorithms are used to optimize decisions in an environment. Examples of rein-
forcement algorithms include Q-learning, Markov decision processes, and Monte Carlo
tree search.
Deep learning and neural networks have become the go-to solutions for many chal-
lenging industrial settings. Deep learning is a subset of machine learning, which seeks
to replicate the way humans absorb and process information and data. Neural net-
works are a form of deep learning that imitate the structure and functions of the
236 S. P. Sreekala et al.
human brain. In industrial settings, deep learning algorithms can be used to catego-
rize and classify large amounts of data, detect patterns, and make predictions and
generate insights [7]. Neural networks have the ability to automate and streamline
decision-making processes, and can be used to improve operational efficiency and re-
duce costs. In some industrial settings, deep learning algorithms can be used to detect
anomalies or patterns with high accuracy as compared to traditional methods. By
feeding these algorithms with large datasets, the algorithms can be trained to identify
features, and even small changes or instances that would otherwise be undetected.
They can also be used to detect objects in images, recognize speech, and model dy-
namic systems. In terms of predictive maintenance, neural networks can be used to
detect subtle anomalies or small changes in the parameters of a system over time.
The neural network will then be able to predict when a system is either likely to fail
or requires maintenance. Deep learning and neural networks have shown some
promise in industrial settings [8]. By combining these technologies with traditional
methods, it is possible to automate and streamline decision-making processes, im-
prove operational efficiency, and reduce costs.
It have been used in several industries such as transportation, retail, and manufactur-
ing. These applications help streamline the supply chain process by reducing paper-
work, increasing efficiency, and lowering costs. In transportation, for example, AI can
help optimize driver routes, reduce unnecessary stops, and identify optimal shipping
methods. AI can also create efficient scheduling to maximize delivery time and reduce
unplanned costs. In retail, AI can be used to automate the supply chain process for
stocking shelves and warehousing inventory [18]. AI can also analyze customer data
240 S. P. Sreekala et al.
to predict trends and recommend items that customers may be more likely to pur-
chase. In manufacturing, AI can be used to automate production scheduling and in-
ventory management. AI can also be used to analyze variables such as cost, lead time,
and quality to optimize production processes and minimize waste. AI applications in
supply chain management can improve the efficiency and effectiveness of the entire
process, helping organizations save time and money.
Statements made on social media and other digital media platforms are an excellent
source of customer data for sentiment analysis. Customer sentiment tracking is the
systematic collection, analysis, and reporting of customer opinions over time. Compa-
nies use customer sentiment tracking to improve their customer experience and to
11 A survey of AI in industry 241
understand how their customers feel about their products and services. This data can
be collected and analyzed using sentiment analysis techniques, such as NLP. Compa-
nies monitor both qualitative (e.g., surveys, customer reviews, and feedback) and
quantitative (e.g., social media posts and sentiment scores) sources to track customer
sentiment. The data is then analyzed in order to identify trends and underlying prob-
lems in customer satisfaction [21].
Fraud detection and prevention using AI algorithms is the process of utilizing Ma-
chine Learning, Deep Learning, and other AI algorithms to automatically detect and
respond to fraud and other malicious activities. AI algorithms can also be used to de-
tect potential threats, such as content manipulation, and cybercrime. By detecting
fraud in near real time, organizations can reduce their chances of costly and damag-
ing fraud. AI-enabled fraud prevention systems can be used to continuously monitor
millions of data points to identify patterns in transactions that may be indicative of
fraudulent activity. AI algorithms can also be used to classify such patterns into risk
ratings, and alert organizations of potential risks [24].
Algorithmic trading and AI-driven investment strategies are a set of rules and algorithms
used to make financial decisions. Algorithmic trading and AI-driven investment strategies
are used to trade stocks and other securities in real time based on various market condi-
tions such as news, market sentiment, and market trends. AI-driven investment strategies
also allow for the automation of all trading decisions and the optimization of portfolio
management. This can help to reduce costs and improve returns. Additionally, AI-driven
investment strategies can leverage complex algorithms and data analysis to identify pat-
terns, detect anomalies, and make decisions regarding investments.
11 A survey of AI in industry 243
Medical imaging and diagnostics with AI assistance is a new field of medical technology
that utilizes AI to help analyze images and provide rapid, accurate diagnoses. AI-assisted
systems use advanced algorithms to process and analyze medical images taken by imag-
ing devices such as ultrasound, magnetic resonance imaging (MRI), X-rays, and computed
tomography (CT) scans. By combining knowledge from large datasets, AI can quickly
identify and categorize objects in an image, such as tumors. Medical imaging and diagnos-
tics with AI assistance can help to reduce the time and cost associated with diagnosing,
monitoring, and managing medical conditions by quickly and accurately analyzing and
interpreting medical imaging and other data. Additionally, AI can be used to reduce false-
positive results and eliminate human error, helping to make diagnoses more accurate.
AI-enabled drug discovery and personalized medicine are two emerging areas of
healthcare that are revolutionizing the way diseases are treated. AI-enabled drug dis-
covery uses data and machine learning to identify and test new potential drugs that
can be used to treat diseases. It can also be used to form clinical decision-making and
help scientists understand why certain diseases occur and how they can be treated.
Personalized medicine is a practice that customizes treatments based on a patient’s
specific needs and genetic markers. By taking into account individual traits, such as
genetic susceptibility or responses to medications, personalized medicine allows
physicians to tailor treatments that are specifically tailored to a patient’s unique
needs. This helps reduce the risk of taking an ineffective or potentially dangerous
treatment and allows physicians to make more informed treatment decisions.
244 S. P. Sreekala et al.
Energy demand forecasting and load optimization with AI is the process of predicting
the energy demand in a particular area and then optimizing the energy supply to
meet that demand using AI. AI models can use various data sources such as historical
data of energy use, temperature, population, economic activity, etc. to make accurate
forecasts of energy demand. AI models can then use complex optimization algorithms
to optimize the energy supply for maximum efficiency, reliability, and cost savings. AI
can track real-time changes and adjust the energy supply accordingly. This enables
organizations to reduce their energy costs, increase energy efficiency, and reduce
their carbon footprint.
11 A survey of AI in industry 245
Smart grid management and energy efficiency using AI is the use of advanced AI to
optimize the efficiency of energy utilization in smart grids. Smart grid management is
a technology-enabled system that enables improved communication between electric-
ity customers and suppliers. AI helps optimize energy usage in smart grids by evaluat-
ing customer usage data in order to identify and act upon opportunities to conserve
energy or reduce operational costs. AI can also enable energy suppliers to optimize
the electricity resources allocated to customers with greater accuracy, helping to re-
duce electricity costs for both consumers and suppliers, and ensuring that electricity
is utilized more efficiently on the grid [27]. AI can also identify potential issues in the
energy grid before they happen and suggest solutions to improve reliability and
safety. Additionally, AI can proactively forecast customer usage demands, allowing
electricity suppliers to anticipate needs and plan for future energy needs accordingly.
By leveraging AI, smart grids can optimize energy efficiency, reduce costs, enhance
reliability, and increase customer satisfaction.
Asset maintenance and predictive analytics in the utilities sector is the practice of utiliz-
ing predictive analytics to analyze operational data from assets in the utilities sector to
identify potential maintenance issues before they occur. This practice helps to reduce
downtime and mitigate the risk associated with asset failure, while enabling utilities to
optimize asset performance and ensure reliable service. Predictive analytics combines
data from existing systems and databases, such as a utility’s asset management system
and geographic information systems, with general patterns of failure, environmental con-
ditions, and other related factors. Based on this data-driven approach, predictive analytics
can predict potential maintenance issues before they happen, allowing utility companies
to take proactive action to address maintenance requirements and minimize asset risk
[28]. In addition to early detection and prevention of asset failures, predictive analytics
can also provide valuable insights into asset utilization and performance. By identifying
asset usage patterns, utilities can improve their operational efficiency and ensure more
reliable service. Predictive analytics can also track the performance of assets over time,
providing detailed feedback on asset health and performance, and allowing for optimiza-
tion of maintenance schedules. Predictive analytics can play an important role in ensur-
ing the reliable, safe, and efficient operation of utility assets. It is a versatile tool that can
alert utilities to potential maintenance issues before they occur, allowing for proactive
preventive action. Additionally, predictive analytics can optimize the efficiency and reli-
ability of assets, ultimately leading to improved customer satisfaction and enhanced fi-
nancial performance.
246 S. P. Sreekala et al.
ence [33]. AI-driven ad campaigns can also enable automated bidding on impressions,
ensuring that ads are served to the most relevant target demographic in the most
cost-effective way. Programmatic advertising and AI-driven ad campaigns are impor-
tant for businesses looking to optimize ROI on their marketing efforts [34].
Social media monitoring and sentiment analysis with AI is a process of using AI algo-
rithms to track, classify, analyze, and interpret public sentiment and online conversa-
tions about a particular product, brand, service, person, or organization [35]. For
instance, AI-driven sentiment analysis can help brands identify how key users are in-
teracting with them and how they are strategically positioning themselves on social
media channels. AI can also help businesses better understand the overall sentiment
expressed by users, and which topics and trends are the most important to them. Ad-
ditionally, AI-powered sentiment analysis can be used to detect and respond to cus-
tomer complaints, identify customer service issues, uncover marketing opportunities,
and monitor competitor strategies [36].
References
[1] Pallathadka, H., Ramirez-Asis, E. H., Loli-Poma, T. P., Kaliyaperumal, K., Ventayen, R. J. M., & Naved,
M. (2023). Applications of artificial intelligence in business management, e-commerce and finance.
Materials Today: Proceedings. 80, 2610–2613.
[2] Kumar, S., Lim, W. M., Sivarajah, U., & Kaur, J. (2023). Artificial intelligence and blockchain
integration in business: Trends from a bibliometric-content analysis. Information Systems Frontiers,
25(2), 871–896.
[3] Ribeiro, J., Lima, R., Eckhardt, T., & Paiva, S. (2021). Robotic process automation and artificial
intelligence in industry 4.0 – a literature review. Procedia Computer Science. 181, 51–58.
[4] Burström, T., Parida, V., Lahti, T., & Wincent, J. (2021). AI-enabled business-model innovation and
transformation in industrial ecosystems: A framework, model and outline for further research.
Journal of Business Research. 127, 85–95.
[5] Kaur, D. N., Sahdev, S. L., Sharma, D. M., & Siddiqui, L. (2020). Banking 4.0:‘The influence of artificial
intelligence on the banking industry & how ai is changing the face of modern day banks’.
International Journal of Management, 11(6), 1–9.
[6] Akter, S., Michael, K., Uddin, M. R., McCarthy, G., & Rahman, M. (2022). Transforming business using
digital innovations: The application of AI, blockchain, cloud and data analytics. Annals of Operations
Research, 2022, 1–33.
[7] Chen, H., Li, L., & Chen, Y. (2021). Explore success factors that impact artificial intelligence adoption
on telecom industry in China. Journal of Management Analytics, 8(1), 36–68.
[8] Lee, D., & Yoon, S. N. (2021). Application of artificial intelligence-based technologies in the
healthcare industry: Opportunities and challenges. International Journal of Environmental Research
and Public Health, 18(1), 271.
248 S. P. Sreekala et al.
[9] Paschen, U., Pitt, C., & Kietzmann, J. (2020). Artificial intelligence: Building blocks and an innovation
typology. Business Horizons, 63(2), 147–155.
[10] Zhang, X., Ming, X., Liu, Z., Yin, D., Chen, Z., & Chang, Y. (2019). A reference framework and overall
planning of industrial artificial intelligence (I-AI) for new application scenarios. The International
Journal of Advanced Manufacturing Technology 101, 2367–2389.
[11] Wamba-Taguimdje, S. L., Fosso Wamba, S., Kala Kamdjoug, J. R., & Tchatchouang Wanko,
C. E. (2020). Influence of artificial intelligence (AI) on firm performance: The business value of
AI-based transformation projects. Business Process Management Journal, 26(7), 1893–1924.
[12] Dash, R., McMurtrey, M., Rebman, C., & Kar, U. K. (2019). Application of artificial intelligence in
automation of supply chain management. Journal of Strategic Innovation and Sustainability, 14(3),
43–53.
[13] Tyagi, A. K., Fernandez, T. F., Mishra, S., & Kumari, S. (2021, June). Intelligent automation systems at
the core of industry 4.0. In Intelligent Systems Design and Applications: 20th International
Conference on Intelligent Systems Design and Applications (ISDA 2020) held December 12–15, 2020
(pp. 1–18). Cham: Springer International Publishing.
[14] Chien, C. F., Dauzère-Pérès, S., Huh, W. T., Jang, Y. J., & Morrison, J. R. (2020). Artificial intelligence in
manufacturing and logistics systems: Algorithms, applications, and case studies. International
Journal of Production Research, 58(9), 2730–2731.
[15] Yadav, S. P., & Yadav, S. (2020). Image fusion using hybrid methods in multimodality medical
images. Medical and Biological Engineering and Computing. 58, 669–687.
[16] Irfan, D., Tang, X., Narayan, V., Mall, P. K., Srivastava, S., & Saravanan, V. (2022). Prediction of Quality
Food Sale in Mart using the AI-based TOR Method. Journal of Food Quality, 2022, 1–13.
[17] Yuvaraj, N., Srihari, K., Dhiman, G., Somasundaram, K., Sharma, A., Rajeskannan, S. M. G. S. M. A., . . .
& Masud, M. (2021). Nature-inspired-based approach for automated cyberbullying classification on
multimedia social networking. Mathematical Problems in Engineering. 2021, 1–12.
[18] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). Energy‐efficient resource allocation
model for device‐to‐device communication in 5G wireless personal area networks. International
Journal of Communication Systems, 2023, 1–14.
[19] Yadav, S. P., Zaidi, S., Mishra, A., & Yadav, V. (2022). Survey on machine learning in speech emotion
recognition and vision systems using a recurrent neural network (RNN). Archives of Computational
Methods in Engineering, 29(3), 1753–1770.
[20] Rajalakshmi, M., Saravanan, V., Arunprasad, V., Romero, C. T., Khalaf, O. I., & Karthik, C. (2022).
Machine learning for modeling and control of industrial clarifier process. Intelligent Automation
and Soft Computing, 32(1), 339–359.
[21] Sivaram, M., Porkodi, V., Mohammed, A. S., Manikandan, V., & Yuvaraj, N. (2019). Retransmission
DBTMA protocol with fast retransmission strategy to improve the performance of MANETs. IEEE
Access. 7, 85098–85109.
[22] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. L.-R. U. B. I. 2023. An efficient load-based
resource utilization algorithm for bi-partite scatternet in wireless personal area networks.
International Journal of Communication Systems, 36(4), 1–17.
[23] Yadav, S. P., Mahato, D. P., & Linh, N. T. D. (Eds.). (2020). Distributed artificial intelligence: A modern
approach. Boca Raton: CRC Press.
[24] Saravanan, V., & Neeraja, A. (2013, January). Security issues in computer networks and
stegnography. In 2013 7th International Conference on Intelligent Systems and Control (ISCO)
(pp. 363–366). IEEE.
[25] Kannan, S., Dhiman, G., Natarajan, Y., Sharma, A., Mohanty, S. N., Soni, M., . . . & Gheisari, M. (2021).
Ubiquitous vehicular ad-hoc network computing using deep neural network with IoT-based bat
agents for traffic management. Electronics, 10(7), 785.
11 A survey of AI in industry 249
[26] Logeshwaran, J., Kiruthiga, T., Kannadasan, R., Vijayaraja, L., Alqahtani, A., Alqahtani, N., & Alsulami,
A. A. (2023). Smart load-based resource optimization model to enhance the performance of device-
to-device communication in 5G-WPAN. Electronics, 12(8), 1821.
[27] Yadav, S. P. (2021). Emotion recognition model based on facial expressions. Multimedia Tools and
Applications, 80(17), 26357–26379.
[28] Narayanan, K. L., Krishnan, R. S., Robinson, Y. H., Julie, E. G., Vimal, S., Saravanan, V., & Kaliappan,
M. (2022). Banana plant disease classification using hybrid convolutional neural network.
Computational Intelligence and Neuroscience, 2022, 1–13.
[29] Sivaram, M., Porkodi, V., Mohammed, A. S., Manikandan, V., & Yuvaraj, N. (2019). Retransmission
DBTMA protocol with fast retransmission strategy to improve the performance of manets. IEEE
Access. 7, 85098–85109.
[30] Ramesh, G., Logeshwaran, J., & Kumar, A. P. (2023). The smart network management automation
algorithm for administration of reliable 5G communication networks. Wireless Communications and
Mobile Computing, 2023, 1–13.
[31] Gupta, R., Tanwar, S., Al-Turjman, F., Italiya, P., Nauman, A., & Kim, S. W. (2020). Smart contract
privacy protection using AI in cyber-physical systems: Tools, techniques and challenges. IEEE Access.
8, 24746–24772.
[32] Sjödin, D., Parida, V., Palmié, M., & Wincent, J. (2021). How AI capabilities enable business model
innovation: Scaling AI through co-evolutionary processes and feedback loops. Journal of Business
Research. 134, 574–587.
[33] Peres, R. S., Jia, X., Lee, J., Sun, K., Colombo, A. W., & Barata, J. (2020). Industrial artificial intelligence
in industry 4.0-systematic review, challenges and outlook. IEEE Access. 8, 220121–220139.
[34] Loureiro, S. M. C., Guerreiro, J., & Tussyadiah, I. (2021). Artificial intelligence in business: State of the
art and future research agenda. Journal of Business Research. 129, 911–926.
[35] Di Vaio, A., Palladino, R., Hassan, R., & Escobar, O. (2020). Artificial intelligence and business models
in the sustainable development goals perspective: A systematic literature review. Journal of
Business Research. 121, 283–314.
[36] Cubric, M. (2020). Drivers, barriers and social considerations for AI adoption in business and
management: A tertiary study. Technology in Society. 62, 101257.
S. Revathy, S. P. Sreekala, D. Praveenadevi, S. Rajeshwari
12 The intelligent implications of artificial
intelligence-driven decision-making
in business management
Abstract: The rapid rate at which technology has progressed in the last few decades
has resulted in emerging technological solutions such as artificial intelligence (AI). AI
is increasingly being used in business management decision-making processes to
make more informed, better-informed, and more intelligent decisions. It is important
to consider the intelligent implications of AI-driven decision-making in business man-
agement in order to ensure that the decisions made are not only effective but are also
ethical, responsible, and beneficial to all stakeholders involved. AI has enabled deci-
sion makers to access and analyze large amounts of data quickly and accurately. This
can aid decision-makers in making better-informed decisions as well as reducing
costs and saving time. AI can also assist decision-makers in identifying important pat-
terns or trends within datasets and in finding solutions to problems that might have
otherwise been overlooked. Additionally, AI can also help to identify potential oppor-
tunities to improve the efficiency of existing processes and systems.
https://doi.org/10.1515/9783111323749-012
252 S. Revathy et al.
Benefits:
– Increased efficiency: AI systems make decisions quickly and accurately, helping
businesses achieve greater cost and time efficiency.
– Improved accuracy: AI systems are capable of gathering, analyzing and interpret-
ing data more accurately than humans can on their own, leading to more reliable
decisions.
254 S. Revathy et al.
Challenges:
– Cost: Implementing AI systems can be quite expensive and time-consuming, espe-
cially in terms of data gathering, training, and maintenance.
– Complexity: AI systems are complex and hard to manage, as they involve multiple
algorithms and data sources.
– Lack of explainability: AI systems are so complex sometimes that it can be diffi-
cult to explain their decisions and outcomes, which has become a focus of re-
search for many companies.
– Human involvement: AI systems still require human input and guidance to en-
sure accurate and ethical results, with the potential for bias in decision-making if
the correct protocols are not in place.
from one of the multiple groups or organizations. AI and its related technologies can
provide valuable data and analysis that can inform human decision-makers, eliminat-
ing bias and increasing accuracy. AI can help in uncovering patterns and associations
that may not be apparent to human decision-makers or predict the probability of fu-
ture events based on current trends and data. This integration increases the effective-
ness of decision-making processes and reduces decision-making errors.
AI-supported market analysis and trend forecasting is a way to use AI in order to ana-
lyze the market and identify trends. AI technology can be used to mine the data from
various sources, and find patterns and correlations. This is helpful for businesses to
make better decisions and have an insight into what will happen in the future [9]. AI-
supported market analysis and trend forecasting can also be used to identify weak
points in the market and opportunities for investment. With AI-assisted market analy-
sis, businesses can save time, increase efficiency, and make smarter decisions.
models can also be used to anticipate market changes and consumer behavior. By
using predictive analytics, businesses can make better informed decisions that can
lead to improved outcomes.
The automation of production planning and scheduling with AI involves the use of AI
technologies for managing the planning and scheduling of production activities in
manufacturing and industrial operations. AI-based production planning and schedul-
ing systems can automate a wide range of tasks such as resource optimization, capac-
ity planning, job sequencing, and task assignment. Not only this, production planners
and schedulers can use AI-driven technologies to improve the accuracy and efficiency
of their operations [14]. With advanced algorithms, these systems can make complex
decisions and recommendations about how to improve the production process. AI-
based production planning and scheduling systems also make it easier to coordinate
258 S. Revathy et al.
tasks across the entire value chain, enabling more efficient utilization of resources
and better decision-making.
Pricing optimization and dynamic pricing with AI are strategies that businesses can
use to maximize their profits. With these strategies, businesses can use AI to deter-
mine the optimal price for each product they sell by considering the pricing history
and the current market environment. AI can take into account customer preferences,
the competitive landscape, and changes in demand patterns to automatically tweak
prices accordingly, ensuring that the company is always in line with what the market
will bear [18]. AI-powered dynamic pricing also allows companies to adjust prices
quickly and accurately in response to customer activity, ensuring that they are always
ahead of the competition and maximizing their profits.
AI-based financial forecasting and risk assessment are built on ML algorithms that
are able to look at an organization’s historical data and predict changes in market
conditions, as well as assess potential risks. For example, an ML algorithm can be
260 S. Revathy et al.
used to look at past financial data and determine the likelihood of a future downturn
in the market or the probability of a particular investment turning a profit.
AI-based financial forecasting and risk assessment are useful as they can help
organizations identify potential risks or opportunities before they happen and
make more informed decisions. They also offer organizations a more detailed and
accurate analysis of their financial situation. AI-based systems can be used to pro-
vide evaluations of risk, identify potential market trends, provide an early warn-
ing system of future financial changes, and even suggest optimal allocation of its
resources. In this way, businesses can make better decisions and improve their fi-
nancial performance.
AI-assisted candidate screening and recruitment processes are the use of AI and ML
technologies to assist with the searching, vetting, and selection of job applicants. AI-
based systems are being used to sort through resumes, identify candidates with skills
and experience that match the job requirements, and provide shortlisted candidates
for interview [24]. AI can also be used to predict the potential success or failure of
candidates through automated scoring methods and to assist with the scheduling of
interviews. It can also be used to retrieve data from potential applicants, such as
references, education records, and other relevant information. AI-assisted candidate
screening and recruitment processes can be used to save time and money, while im-
proving the accuracy and reliability of the results.
Performance evaluation and talent management with AI is a system that uses AI algo-
rithms to automate the process of evaluating employees’ performance and managing
talent. This system uses analytical solutions to determine employee potential, skills,
opportunities for improvement, and best way to allocate resources to achieve organi-
262 S. Revathy et al.
zational goals. The system typically includes an AI-driven 360-degree feedback assess-
ment which, unlike traditional feedback methods, is designed to anonymously evalu-
ate the employee’s performance on multiple levels. AI can also be used to identify
skills gaps and match employees to roles or development opportunities. Additionally,
AI can be used to create custom development plans for each employee based on their
current skills and future goals [25]. AI can provide data-driven insights into organiza-
tional culture and overall performance, allowing organizations to ensure alignment
with their values, and employees to better understand organizational expectations.
Employee engagement and sentiment analysis using AI refers to the use of AI techni-
ques to measure and analyze employee engagement and sentiment. AI technologies
such as NLP and ML can be used to automatically analyze employee feedback in written
and spoken forms and to identify sentiment and engagement levels. These analyses can
provide valuable insights for organizations to better understand their employees and
optimize their employee experience efforts. AI technologies can also be used to create
automated employee sentiment surveys and identify areas within an organization that
can be improved to increase employee engagement and satisfaction.
data about customers to better understand, engage with, and target them. Predictive
modeling is a type of ML that uses historical customer data to build models to predict
future customer behavior. These models can be used to identify the most promising
customer segments, understand customer preferences, and target and personalize
marketing messages to improve conversion and retention [27]. By leveraging AI and
ML to analyze customer data, businesses can more effectively target potential custom-
ers, deliver effective campaigns, and improve customer service. This can result in in-
creased sales, higher customer satisfaction, and better customer retention.
Job roles and workforce dynamics are constantly changing and evolving due to the
impact of technology. Many industries are seeing increased automation and techno-
logical advancements, leading to new and different roles that may not have even ex-
isted before. As technology advances, traditional roles become obsolete and people
may have to learn new skills and take up new roles in order to remain relevant in
their industry. This shift from manual labor-based roles to more technical ones can
have a huge impact on a workforce’s dynamics and should be approached with con-
siderable thought and planning [31]. A workforce must stay up-to-date on the latest
trends and be willing to learn new skills if they want to remain competitive in the
modern economy. Employers must be aware of the implications of technology on job
roles and workforce dynamics and become proactive in anticipating and training for
the changes. It is also important to address the issue of displacement of existing work-
ers who may not be able to develop new skills or who may become obsolete in their
current roles [32]. Employers can take measures to ensure job security for existing
workers through retraining initiatives and creating new roles that may not have been
possible before. Technological advancements are having a huge impact on job roles
and workforce dynamics, and the changes are evolving rapidly. Companies must be
aware of the changes and be prepared to adjust their strategies accordingly in order
to remain successful.
12 The intelligent implications of artificial 265
Regulatory and legal considerations are important when adopting AI, and can still
vary depending on the industry, country, and business. The most common legal con-
siderations include compliance with laws and regulations applicable to the develop-
ment, use, and deployment of AI, intellectual property protection, data privacy and
security, contracts to regulate the relationship between the parties involved in the AI
initiative, and an assessment of potential personal injury and liability claims should
something unexpectedly go wrong due to the AI system. Regulatory considerations in-
volve the potential implications of regulatory requirements on the development, use,
and deployment of AI. This involves understanding how the existing laws, regulations,
and initiatives may overlap with each other in relation to AI technology, including
sector-specific regulatory requirements across borders [34]. When implementing AI
technology, organizations should ensure that privacy concerns and other applicable
laws and regulations are taken into account. This includes compliance with consumer
protection laws, data privacy laws, and labor and employment regulations. Organiza-
tions should also engage in risk management to identify potential risks associated
with implementing AI technology. This should involve considering compliance risks,
the potential for bias, data security and privacy risks, and the potential for legal ac-
tion related to the AI system.
266 S. Revathy et al.
References
[1] Novak, A., Bennett, D., & Kliestik, T. (2021). Product decision-making information systems, real-time
sensor networks, and artificial intelligence-driven big data analytics in sustainable industry 4.0.
Economics, Management and Financial Markets, 16(2), 62–72.
[2] Rajagopal, N. K., Qureshi, N. I., Durga, S., Ramirez Asis, E. H., Huerta Soto, R. M., Gupta, S. K., &
Deepak, S. (2022). Future of business culture: An artificial intelligence-driven digital framework for
organization decision-making process. Complexity, 2022, 14. Article ID 7796507.
[3] Andronie, M., Lăzăroiu, G., Iatagan, M., Uță, C., Ștefănescu, R., & Cocoșatu, M. (2021). Artificial
intelligence-based decision-making algorithms, internet of things sensing networks, and deep
learning-assisted smart process management in cyber-physical production systems. Electronics,
10(20), 2497.
[4] Adams, D., & Krulicky, T. (2021). Artificial intelligence-driven big data analytics, real-time sensor
networks, and product decision-making information systems in sustainable manufacturing internet
of things. Economics, Management and Financial Markets, 16(3), 81–93.
[5] Schrettenbrunnner, M. B. (2020). Artificial-intelligence-driven management. IEEE Engineering
Management Review, 48(2), 15–19.
[6] Dubey, R., Bryde, D. J., Dwivedi, Y. K., Graham, G., & Foropon, C. (2022). Impact of artificial
intelligence-driven big data analytics culture on agility and resilience in humanitarian supply chain:
A practice-based view. International Journal of Production Economics, 250, 108618.
[7] Wong, L. W., Tan, G. W. H., Ooi, K. B., Lin, B., & Dwivedi, Y. K. (2022). Artificial intelligence-driven risk
management for enhancing supply chain agility: A deep-learning-based dual-stage PLS-SEM-ANN
analysis. International Journal of Production Research, 2022, 1–21.
[8] Belhadi, A., Mani, V., Kamble, S. S., Khan, S. A. R., & Verma, S. (2021). Artificial intelligence-driven
innovation for enhancing supply chain resilience and performance under the effect of supply chain
dynamism: An empirical investigation. Annals of Operations Research, 2021, 1–26.
12 The intelligent implications of artificial 267
[9] Zohuri, B., & Rahmani, F. M. (2023). Artificial intelligence driven resiliency with machine learning and
deep learning components. Japan Journal of Research, 1(1), 1–7.
[10] Dey, P. K., Chowdhury, S., Abadie, A., Vann Yaroson, E., & Sarkar, S. (2023). Artificial intelligence-
driven supply chain resilience in Vietnamese manufacturing small-and medium-sized enterprises.
International Journal of Production Research, 2023, 1–40.
[11] Nica, E., & Stehel, V. (2021). Internet of things sensing networks, artificial intelligence-based
decision-making algorithms, and real-time process monitoring in sustainable industry 4.0. Journal
of Self-Governance and Management Economics, 9(3), 35–47.
[12] Pietronudo, M. C., Croidieu, G., & Schiavone, F. (2022). A solution looking for problems? A systematic
literature review of the rationalizing influence of artificial intelligence on decision-making in
innovation management. Technological Forecasting and Social Change, 182, 121828.
[13] Faqihi, A., & Miah, S. J. (2023). Artificial intelligence-driven talent management system: Exploring the
risks and options for constructing a theoretical foundation. Journal of Risk and Financial
Management, 16(1), 31.
[14] Nica, E., Stan, C. I., Luțan, A. G., & Oașa, R. Ș. (2021). Internet of things-based real-time production
logistics, sustainable industrial value creation, and artificial intelligence-driven big data analytics in
cyber-physical smart manufacturing systems. Economics, Management, and Financial Markets, 16(1),
52–63.
[15] Lawrence, J., & Durana, P. (2021). Artificial intelligence-driven big data analytics, predictive
maintenance systems, and internet of things based real-time production logistics in sustainable
industry 4.0 wireless networks. Journal of Self-Governance and Management Economics, 9(4), 1–12.
[16] Pujahari, R. M., Yadav, S. P., & Khan, R. (2022). Intelligent farming system through weather forecast
support and crop production. In: Mohammad Ayoub Khan, Rijwan Khan & Mohammad Aslam Ansari
Application of Machine Learning in Agriculture (pp. 113–130). United States: Academic Press.
[17] Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J., . . . &
Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey
wolf optimization in wireless sensor networks. Security and Communication Networks, 2022, 1–13.
[18] Yuvaraj, N., Srihari, K., Chandragandhi, S., Raja, R. A., Dhiman, G., & Kaur, A. (2021). Analysis of
protein-ligand interactions of sars-cov-2 against selective drug using deep neural networks. Big
Data Mining and Analytics, 4(2), 76–83.
[19] Logeshwaran, J., Kiruthiga, T., Kannadasan, R., Vijayaraja, L., Alqahtani, A., Alqahtani, N., & Alsulami,
A. A. (2023). Smart load-based resource optimization model to enhance the performance of device-
to-device communication in 5G-PAN. Electronics, 12(8), 1821.
[20] Singh, P. K., Chauhan, S. S., Sharma, A., Prakash, S., & Singh, Y. (2023). Prediction of higher heating
values based on imminent analysis by using regression analysis and artificial neural network for
bioenergy resources. In Proceedings of the Institution of Mechanical Engineers, Part E: Journal of
Process Mechanical Engineering, 09544089231175046.
[21] Saravanan, V., & Mohan Raj, V. (2016). Maximizing QoS by cooperative vertical and horizontal
handoff for tightly coupled WIMAX/WLAN overlay networks. Cluster Computing, 19(3), 1619–1633.
[22] Mojjada, R. K., Yadav, A., Prabhu, A. V., & Natarajan, Y. (2020). Machine learning models for covid-19
future forecasting. Materials Today: Proceedings 2020, 1–5.
[23] Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model
for efficient resource allocation in 5G broadband communication networks. ICTACT Journal on Soft
Computing, 13(3), 2986–2994.
[24] Yadav, S. P. (2020). Vision-based detection, tracking, and classification of vehicles. IEIE Transactions
on Smart Processing and Computing, 9(6), 427–434.
[25] Sumathi, A., & Saravanan, V. (2015). Bandwidth based vertical handoff for tightly coupled WIMAX/
WLAN overlay networks 74(10), 560–566.
268 S. Revathy et al.
[26] Manikandan, V., & Yuvaraj, N. (2019, May). Advanced expert system using particle swarm
optimization based adaptive network based fuzzy inference system to diagnose the physical
constitution of human body. In Emerging Technologies in Computer Engineering: Microservices in
Big Data Analytics: Second International Conference, ICETCE 2019, Jaipur, India, February 1–2, 2019,
Revised Selected Papers (Vol. 985, pp. 349). Springer.
[27] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). Energy‐efficient resource allocation
model for device‐to‐device communication in 5G wireless personal area networks. International
Journal of Communication Systems, 2023,1–14.
[28] Yadav, S. P., Mahato, D. P., & Linh, N. T. D. (Eds.). (2020). Distributed artificial intelligence: A modern
approach. Boca Raton: CRC Press.
[29] Rajalakshmi, M., Saravanan, V., Arunprasad, V., Romero, C. T., Khalaf, O. I., & Karthik, C. (2022).
Machine learning for modeling and control of industrial clarifier process. Intelligent Automation
and Soft Computing, 32(1) 339–359.
[30] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. L-RUBI (2023). An efficient load-based resource
utilization algorithm for bi-partite scatternet in wireless personal area networks. International
Journal of Communication Systems, 36(4), 1–17.
[31] Yuvaraj, N., Raja, R. A., Kousik, N. V., Johri, P., & Diván, M. J. (2020). Analysis on the prediction of
central line-associated bloodstream infections (CLABSI) using deep neural network classification. In
Computational Intelligence and its Applications in Healthcare (pp. 229–244). India: Academic Press.
[32] Belhadi, A., Kamble, S., Fosso Wamba, S., & Queiroz, M. M. (2022). Building supply-chain resilience:
An artificial intelligence-based technique and decision-making framework. International Journal of
Production Research, 60(14), 4487–4507.
[33] Ganapathy, A., Redwanuzzaman, M., Rahaman, M. M., & Khan, W. (2020). Artificial intelligence
driven crypto currencies. Global Disclosure of Economics and Business, 9(2), 107–118.
[34] Rodgers, W., Yeung, F., Odindo, C., & Degbey, W. Y. (2021). Artificial intelligence-driven music
biometrics influencing customers’ retail buying behavior. Journal of Business Research, 126,
401–414.
[35] Zohuri, B., & Moghaddam, M. (2020). From business intelligence to artificial intelligence. Journal of
Material Sciences and Manufacturing Research. SRC/JMSMR/102 Page, 3.
P. Krishna Priya, Karthik Rudra, Deepak Sai Kandula, G. Charan Teja,
K. Venkata Koushik Reddy
13 An innovative analysis of AI-powered
automation techniques for business
management
Abstract: Artificial intelligence (AI) is rapidly changing the way businesses manage
their operations. AI-powered automation techniques are becoming an increasingly im-
portant tool for enterprise success. Automation helps businesses save time and money
by eliminating or reducing manual tasks to streamline processes. Automation also ena-
bles decision-making with greater accuracy and speed. AI-powered automation can im-
prove efficiency in business operations by automating repetitive tasks and saving time.
Automated systems can also make personal connections with customers by recognizing
their preferences and responding accordingly. AI tools can even predict trends and rec-
ommend solutions. Automation also helps improve customer experience by providing
improved customer service and increased accuracy when interacting with customers.
https://doi.org/10.1515/9783111323749-013
270 P. Krishna Priya et al.
Automated data collection and analysis using AI algorithms is the process of using AI
techniques to gather, process, and analyze data. It is used to identify patterns and
draw insights from large and complex data sets. The algorithm-driven approach is
based on self-learning, data-driven models, which can be trained to process massive
amounts of data with higher accuracy and in lesser time than traditional methods.
This allows businesses to quickly identify opportunities, trends, and anomalies in
data, leading to improved decision-making and business outcome. AI techniques such
as machine learning, NLP, computer vision, and deep learning are used to collect,
sort, and analyze data on a large scale, discovering patterns and providing insights
that can inform decisions and lead to operational improvements.
AI chatbots and virtual assistants provide automated customer support. They are es-
sentially computer programs that simulate human conversations, assisting customers
with questions, giving advice, and offering helpful solutions. Customers can interact
with chatbots through text or voice commands. Chatbots and virtual assistants are
great tools for providing customers with quick, safe, and cost-effective service. They
can reduce wait times and decrease the cost of live customer support. Additionally, as
13 An innovative analysis of AI-powered automation techniques 275
AI solutions become more sophisticated, they can identify customer needs and proac-
tively provide individualized customer service [12]. AI chatbots and virtual assistants
are becoming increasingly popular in customer service, and businesses across various
industries are already implementing them to improve customer experiences.
Automated ticket routing and issue resolution using AI is a new approach to customer
service that uses AI to read customer messages and route them to the most appropri-
ate agent. AI is also used to identify the root cause of customer issues and present a
suggested resolution to the agent. This helps customer service agents provide more
consistent and accurate customer support. AI also helps support teams by bringing in
contextual information to agents, such as customer history, custom field values, and
knowledge from third-party sources [14]. This allows agents to provide faster and
more contextual support, reducing customer frustration. Ultimately, this will help
teams reduce ticket resolution time, improve customer satisfaction, and reduce cus-
tomer support costs.
276 P. Krishna Priya et al.
AI-based fraud detection and risk assessment is a technology that uses AI to detect
and manage potential fraud and risk situations. It can scan large volumes of data and
quickly identify suspicious activity or transactions that could be indicative of fraud.
By applying machine learning algorithms such as supervised learning, unsupervised
learning, and neural networks, they can establish patterns in the data, identify anom-
alies, and develop new rules for detecting fraud. AI-driven technology allows compa-
nies to gain more insights from their data and to more efficiently respond to fraud
threats as they appear [16]. AI-based fraud detection and risk assessment also enables
13 An innovative analysis of AI-powered automation techniques 277
financial institutions to quickly respond to threats and protects their customers from
potential losses due to fraudulent activity.
AI-driven lead generation and prospect targeting is a process in which AI and machine
learning algorithms are used to locate and identify potential customers and opportuni-
ties for marketing, sales, and other business pursuits. AI-driven lead generation and
prospect targeting technology helps marketers and sales teams understand customer
behavior and preferences, thus offering more accurate insights when it comes to target-
ing the right customer with the right product or service. It can collect and analyze data
from various sources –website visits, social media, and email marketing – in order to
identify potential leads as well as send out tailored communications that are more
likely to lead to conversions.
278 P. Krishna Priya et al.
AI-based demand forecasting and inventory optimization are two techniques that
businesses use to improve their bottom lines. Demand forecasting uses historical data
and predictive analytics to anticipate customer demand, allowing companies to better
plan for future supply needs. Inventory optimization uses AI-based algorithms and
AI-driven data analysis to determine which items in inventory should be kept in stock
and which ones should be sold off. Both techniques help businesses adjust their order-
ing and stocking decisions to ensure that they have the right amount of product avail-
able to meet customers’ needs. By doing so, businesses can reduce waste and save
money by avoiding excess inventory or running out of stock. Having an accurate de-
mand forecast and inventory optimization strategy also helps businesses better serve
their customers, keeping satisfaction levels high.
ability of the suppliers, pricing, delivery times, and more. The use of automation can
also provide visibility into historical supplier data in order to make the best decisions.
will be used in the real world and what potential implications this may have for indi-
viduals, groups, and society at large [30]. This means thinking about potential biases,
making sure datasets are representative and accurate, and considering privacy and
security issues. Additionally, responsible AI automation involves taking time to think
through and clearly articulate the rules that define how the AI system should operate,
as well as providing ways for users to provide feedback or request changes in the
event that unforeseen issues arise.
References
[1] Dam, H. K., Tran, T., Grundy, J., Ghose, A., & Kamei, Y. (2019, May). Towards effective AI-powered
agile project management. In 2019 IEEE/ACM 41st International Conference on Software
Engineering: New Ideas and Emerging Results (ICSE-NIER) (pp. 41–44). IEEE.
[2] Militano, L., Zafeiropoulos, A., Fotopoulou, E., Bruschi, R., Lombardo, C., Edmonds, A., &
Papavassiliou, S. (2021, December). AI-powered infrastructures for intelligence and automation in
beyond-5g systems. In 2021 IEEE Globecom Workshops (GC Wkshps) (pp. 1–6). IEEE.
[3] Kumar, M., Raut, R. D., Mangla, S. K., Ferraris, A., & Choubey, V. K. (2022). The adoption of artificial
intelligence powered workforce management for effective revenue growth of micro, small, and
medium scale enterprises (MSMEs). Production Planning and Control, 2022, 1–17.
[4] Beer Mohamed, M. I., & Hassan, M. F. (2022, November). Adaptive security for cognitive robotic
process automation in enterprise computing using AI-powered analytical engine. In International
Conference on Artificial Intelligence for Smart Community: AISC 2020, 17–18 December, Universiti
Teknologi Petronas, Malaysia (pp. 825–835). Singapore: Springer Nature Singapore.
[5] Martínez-Rojas, A., Sánchez-Oliva, J., López-Carnicer, J. M., & Jiménez-Ramírez, A. (2021, August).
AIRPA: An architecture to support the execution and maintenance of AI-powered RPA robots. In
Business Process Management: Blockchain and Robotic Process Automation Forum: BPM 2021
Blockchain and RPA Forum, Rome, Italy, September 6–10, 2021, Proceedings (pp. 38–48). Cham::
Springer International Publishing.
[6] Younis, R., & Adel, H. M. (2020, September). Artificial intelligence strategy, creativity-oriented HRM
and knowledge-sharing quality: Empirical analysis of individual and organisational performance of
AI-powered businesses. In The Annual International Conference of The British Academy of
Management (BAM).
[7] Pillai, R., Sivathanu, B., & Dwivedi, Y. K. (2020). Shopping intention at AI-powered automated retail
stores (AIPARS). Journal of Retailing and Consumer Services, 57, 102207.
[8] Kulkarni, P., Bhoj, R., Rajapure, S., Gujar, S., Verma, D., & Ranjan, R. (2021, October). Camelot: AI-
Powered Campaign Management System. In 2021 2nd Global Conference for Advancement in
Technology (GCAT) (pp. 1–6). IEEE.
[9] Recto, R. (2022). AI-powered tools at work; zoned in or zoned out?: A correlational study between AI
and flow experiences in the workplace (Doctoral dissertation, Azusa Pacific University).
[10] Ashri, R. (2019). The AI-powered workplace: How artificial intelligence, data, and messaging
platforms are defining the future of work. India: Apress.
[11] Vijayakumar, H. (2023). Business value impact of AI-Powered Service Operations (AIServiceOps).
Available at SSRN 4396170.
[12] He, A. Z., & Zhang, Y. (2022). AI-powered touch points in the customer journey: A systematic literature
review and research agenda. Journal of Research in Interactive Marketing, ahead-of-print, 1–20.
[13] Brown, N., Brown, A., Degupta, A., Quinn, B., Stringer, D., & Yankov, B. (2021, September). Industry
first AI-powered fully automated safety observation system deployed to global offshore fleet. In SPE
Offshore Europe Conference & Exhibition. OnePetro.
[14] Rosado, L., Faria, P., Gonçalves, J., Silva, E., Vasconcelos, A., Braga, C., . . . & Carlos, C. (2022).
EyesOnTraps: AI-powered mobile-based solution for pest monitoring in viticulture. Sustainability,
14(15), 9729.
[15] Ma, Y., Ping, K., Wu, C., Chen, L., Shi, H., & Chong, D. (2020). Artificial intelligence powered internet
of things and smart public service. Library Hi Tech, 38(1), 165–179.
13 An innovative analysis of AI-powered automation techniques 285
[16] Xin, Y., Yang, K., Chih-Lin, I., Shamsunder, S., Lin, X., & Lai, L. (2023). Guest editorial: AI-powered
Telco Network Automation: 5G evolution and 6G. IEEE Wireless Communications, 30(1), 68–69.
[17] Yadav, S. P., & Yadav, S. (2019). Mathematical implementation of fusion of medical images in
continuous wavelet domain. Journal of Advanced Research in Dynamical and Control System, 10(10),
45–54.
[18] Saravanan, V., & Sumathi, A. (2012). Handoff mobiles with low latency in heterogeneous networks
for seamless mobility: A survey and future directions. European Journal of Scientific Research, 81(3),
417–424.
[19] Yuvaraj, N., Praghash, K., Raja, R. A., & Karthikeyan, T. (2022). An investigation of garbage disposal
electric vehicles (GDEVs) integrated with deep neural networking (DNN) and intelligent
transportation system (ITS) in smart city management system (SCMS). Wireless Personal
Communications, 123(2), 1733–1752.
[20] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). Energy‐efficient resource allocation
model for device‐to‐device communication in 5G wireless personal area networks. International
Journal of Communication Systems, 2023, 1–14.
[21] Yadav, S. P., & Yadav, S. (2018). Fusion of medical images in wavelet domain: A discrete
mathematical model. Ingeniería Solidaria, 14(25), 1–11.
[22] Sumathi, A., & Saravanan, V. (2015). Bandwidth based vertical handoff for tightly coupled WIMAX/
WLAN overlay networks, Journal of Scientific & Industrial Research, 74(10), 560–566.
[23] Natarajan, Y., Kannan, S., & Mohanty, S. N. (2021). Survey of various statistical numerical and
machine learning ontological models on infectious disease ontology. Data Analytics in
Bioinformatics: A Machine Learning Perspective, 13(1), 431–442.
[24] Logeshwaran, J., Kiruthiga, T., Kannadasan, R., Vijayaraja, L., Alqahtani, A., Alqahtani, N., & Alsulami,
A. A. (2023). Smart load-based resource optimization model to enhance the performance of device-
to-device communication in 5G-WPAN. Electronics, 12(8), 1821.
[25] Yadav, S. P., & Yadav, S. (2019). Fusion of medical images using a wavelet methodology: A survey.
IEIE Transactions on Smart Processing and Computing, 8(4), 265–271.
[26] Subramanian, B., Saravanan, V., Nayak, R. K., Gunasekaran, T., & Hariprasath, S. (2019). Diabetic
retinopathy-feature extraction and classification using adaptive super pixel algorithm. International
Journal of Engineering and Advanced Technology, 9, 618–627.
[27] Shukla, A., Kalnoor, G., Kumar, A., Yuvaraj, N., Manikandan, R., & Ramkumar, M. (2021). Improved
recognition rate of different material category using convolutional neural networks. Materials
Today: Proceedings. 81(2), 947–950.
[28] Sangeetha, R., Logeshwaran, J., Rocher, J., & Lloret, J. (2023). An improved agro deep learning model
for detection of Panama Wilts disease in banana leaves. AgriEngineering, 5(2), 660–679.
[29] Vashisht, V., Pandey, A. K., & Yadav, S. P. (2021). Speech recognition using machine learning. IEIE
Transactions on Smart Processing and Computing, 10(3), 233–239.
[30] Sekaran, S. C., Saravanan, V., RudraKalyanNayak, R., & Shankar, S. S. (2019). Human health and
velocity aware network selection scheme for WLAN/WiMAX integrated networks with QoS.
International Journal of Innovative Technology and Exploring Engineering (IJITEE), 8(12), 198–205.
[31] Yuvaraj, N., Kousik, N. V., Raja, R. A., & Saravanan, M. (2020). Automatic skull-face overlay and
mandible articulation in data science by AIRS-Genetic algorithm. International Journal of Intelligent
Networks, 1, 9–16.
[32] Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model
for efficient resource allocation in 5G broadband communication networks. ICTACT Journal On Soft
Computing, 13(3), 2986–2994.
286 P. Krishna Priya et al.
[33] Jiménez-Ramırez, A. (2021, August). AIRPA: An architecture to support the execution and
maintenance of AI-powered RPA Robots. In Business Process Management: Blockchain and Robotic
Process Automation Forum: BPM 2021 Blockchain and RPA Forum, Rome, Italy, September 6–10,
2021, Proceedings (Vol. 428, pp. 38). Springer Nature.
[34] Doshi, M., & Varghese, A. (2022). Smart agriculture using renewable energy and AI-powered IoT. In:
Ajith Abraham & Sujata Dash. AI, Edge and IoT-based Smart Agriculture (pp. 205–225). United
States: Academic Press.
B. Girimurugan, S. Rajeshwari, S. P. Sreekala, S. Revathy
14 The smart and secured AI-powered
strategies for optimizing processes
in multi-vendor business applications
Abstract: The development of artificial intelligence (AI)-powered processes in multi-
vendor business applications has created significant opportunities to optimize resour-
ces, streamline operations, and reduce costs. However, deploying AI in multi-vendor
business applications presents a unique set of challenges, including the potential for
data privacy and security issues. To ensure the success of AI-enabled processes, organ-
izations should implement best practices for proper data protection and secure AI-pow-
ered strategies. With the proper implementation of data security protocols and usage
strategies, organizations can harness the potential of AI-powered processes in multi-
vendor business applications to drive greater efficiency and cost savings. This includes
creating greater process efficiency, reducing manual labor costs, and improving cus-
tomer experiences. This will also enable organizations to remain competitive in the rap-
idly changing business landscape.
https://doi.org/10.1515/9783111323749-014
288 B. Girimurugan et al.
the organization should establish KPIs to measure the performance of AI-enabled pro-
cesses, and invest in continuing training and education to ensure the AI-powered sys-
tems remain effective and adaptive.
The focus of smart and secured AI-powered strategies is to develop strategies that en-
sure the security, safe development, and responsible use of AI solutions and services.
This includes strategies to protect critical infrastructure, reduce the risk of cyberat-
tacks, ensure ethical practices, protect personal data, and foster trust and accountabil-
ity. It also aims to create strategies that help to address the economic, regulatory, and
ethical challenges that accompany the use of AI and its related technologies [2].
Through these strategies, organizations and businesses can ensure that their AI solu-
tions and services are secure, accountable, and responsible. The ultimate objective is
to develop effective policy frameworks that support trust and innovation and ensure
that AI is used responsibly to benefit society as a whole.
error or require manual entry. This automation can save time, improve accuracy, in-
crease customer satisfaction, and reduce costs. AI technologies such as NLP, predictive
analytics, and ML can also be used to improve forecasting and scheduling, identify
potential process improvements, and recommend process solutions. Additionally,
AI-driven process optimization can enable faster decision-making, analyze big data
for trends and patterns, improve customer segmentation, and even enable predictive cus-
tomer service. By using AI to optimize processes, companies can reduce costs, improve
efficiency, and remain competitive.
Data inconsistency refers to differences in some way when the same data is stored in two
different locations, or across multiple systems. This can cause difficulty in searching and
retrieving data and may lead to incorrect comparisons and inaccurate results [6].
Interoperability challenges refer to the ability of two or more different systems or
networks to communicate, exchange data, and use the information that has been ex-
changed to enable mutual tasks. Poor interoperability between systems can cause delays
in data transfer, inaccurate data exchange, and incomplete integration of applications. It
can also lead to increased transaction costs and a decrease in overall efficiency.
Performance bottlenecks occur when a machine or a system reaches its maximum ca-
pacity and can no longer process data efficiently. This may result in decreased perfor-
mance, slower loading times, and instability due to system overload.
Resource allocation problems occur when resources such as memory, processing
power, and storage capacity become limited. This may lead to higher costs, wasted
time, and hindered productivity. As resources become less available over time, sys-
tems must be designed more efficiently in order to maximize their use. This can be
achieved through careful planning and optimization.
14 The smart and secured AI-powered strategies for optimizing processes 291
predictive analytics offers an efficient and effective way to utilize this data for plan-
ning and resource optimization.
Real-time monitoring and adaptive process control with AI and machine reading (MR)
allow for automation and optimization of an entire process operation. The entire pro-
cess can be monitored in real time using AI-driven analytics, and dynamic process
control algorithms can be implemented to quickly respond to changes in the process
environment and avoid costly production stoppages and downtime. AI-driven analyt-
ics can also contribute to predictive maintenance alerts, preventive maintenance
plans, and improved equipment availability [10]. MR techniques such as Optical Char-
acter Recognition (OCR) can be used to extract useful information from process data,
and to make predictive process control decisions based on this information. This in-
formation can be used to improve product quality, shorten cycles, increase through-
put, and reduce energy consumption. With the help of AI, adaptive process control
can maintain optimal conditions in the process parameters, while allowing the pro-
cess to quickly adjust to new changes or demands.
cious user activity or anomalous behavior and report any threats or anomalies in
near-real time. Security teams can use AI-powered automation to intelligently re-
spond to these threats quickly and appropriately. AI-powered strategies can also be
used to secure enterprise systems, networks, applications, and data – across both on-
premise and cloud-based infrastructures. This results in improved detection power
and better protection against malicious actors. Additionally, AI-driven security sys-
tems can conduct an analysis of data and automate routine tasks such as patching
and vulnerability scanning to improve security and avoid human errors.
ing, and other mechanisms. To ensure a secure multi-vendor system, these authentica-
tion and access control mechanisms must be regularly maintained and updated as
new users, vendors, and resources are added to the system.
Threat detection and anomaly detection using AI-based techniques allow organiza-
tions to identify and prevent cyberattacks with a greater degree of accuracy than tra-
ditional security solutions. AI-based systems leverage data to learn and relearn the
normal pattern of user activity, and can detect deviations from that pattern, alerting
security teams to possible attacks in progress. AI-based systems are powered by large
datasets of past security incidents, known as “ground truth.” This data can be used to
help ML algorithms detect anomalies faster and more efficiently than other methods,
such as signature-based detection [12]. Furthermore, AI-based systems can be tailored
to the particular needs of an organization, allowing it to take into account the unique
characteristics of the organization and its security profile.
accurately, and optimize the utilization of resources such as time, energy, and money.
Enhanced decision-making capabilities and resource utilization can help organiza-
tions make more informed and strategic decisions that can potentially lead to greater
gains in productivity, efficiency, and profitability. Additionally, this approach can
lead to better customer service and satisfaction as well as improved morale among
employees.
Collaboration and partnerships are models for successful implementation of any project
or initiative. The aim of these models is to ensure integrated and effective delivery of
services, resources, and strategies that result in increased satisfaction for multiple stake-
holders, including customers, employees, and communities. Collaboration refers to an in-
tentional relationship between two or more organizations that work together to identify,
develop, and deliver projects and initiatives through a shared vision, strategy, and goals.
Partnerships are similar but not the same as collaborations in that they involve a specific
agreement involving defined roles of both parties to achieve a specific-defined outcome.
Both models involve more than just planning, coordination, and implementation. Success-
ful collaboration and partnership models must also include shared data and resources;
compatible organizational cultures; agreed-upon goals and objectives; a planned budget
and timeline; and a shared commitment to the project’s success. The collaboration and
partnership models for successful implementation should be built on trust, respect, com-
munication, and shared resources [23]. Such partnerships must also be managed by
a process that encourages transparency and open communication, and which pro-
vides accountability and metrics that measure success. These models should also
strive to be continuously improved, as needed, to ensure successful outcomes.
it ensures that the strategy is working and that it will deliver the desired outcomes. It
also helps to identify areas where improvements can be made and updates can be
made to the strategy, if necessary. Continuous monitoring and evaluation of AI-
powered strategies provides the visibility and transparency required to assess the ef-
fectiveness of deployed AI implementations. It also helps detect any anomalies and
potential risks that could be affecting the success of the implementation. In short, con-
tinuous monitoring and evaluation of AI-powered strategies allows businesses to en-
sure that their strategies are working correctly and make any adjustments that may
be needed.
Compliance with data protection and privacy regulations is a critical issue for any or-
ganization that stores, processes, or transmits personal data. It refers to the laws, reg-
ulations, and other requirements that organizations must adhere to in order to
ensure that data security and privacy are properly safeguarded. This includes laws
such as the EU General Data Protection Regulation (GDPR), the International Organi-
zation for Standardization (ISO) 27001, and any government regulations applicable to
data privacy. Data protection and privacy compliance involves crafting and following
clear processes for collecting, storing, modifying, encrypting, securing, and disposing
of data – and ensuring that employees understand their data protection responsibili-
ties. Organizations must also create processes to regularly monitor how data privacy
is being maintained. Adhering to data protection and privacy regulations helps organ-
izations protect the data of their customers, partners, and employees as well as their
own business reputation.
tial bias. In addition, organizations should ensure that their data privacy and security
policies are up-to-date and that they are compliant with applicable laws. Finally, or-
ganizations should ensure that they have a process and mechanism for monitoring
and detecting bias in their models. This includes using a suite of data-quality tools as
well as an AI ethics framework to help guide data collection and usage decisions.
[28]. AI-based predictive analytics can be used to anticipate outcomes or trends, en-
abling organizations to adjust their operations in real time to stay ahead of changes or
challenges. This proactive approach tends to improve operational efficiency, as organi-
zations can adjust in advance to changes in the market or environment. AI technologies
and their applications have had a positive impact on process optimization. Organiza-
tions now have the potential to gain greater insights into their operations and stream-
line their processes more effectively than ever before.
AI governance frameworks are a set of tools, processes, and decisions designed to help
organizations manage the deployment and use of AI technology. These frameworks pro-
vide structure and guidelines to ensure that AI applications are used responsibly and
efficiently. This includes setting parameters for data protection, decision management,
and ethical application of AI technologies. As the use of AI technology has grown, so has
the need for more sophisticated governance frameworks. Multi-vendor applications,
such as those used in multi-cloud environments, often require more complex gover-
nance models. For example, each cloud vendor has its own unique security and compli-
ance requirements, and the combination of multiple vendors can create additional
complexities [29]. In order to address these complexities, a number of AI governance
frameworks have emerged that are designed to provide guidance and structure to
14 The smart and secured AI-powered strategies for optimizing processes 305
References
[1] Xin, Y., Yang, K., Chih-Lin, I., Shamsunder, S., Lin, X., & Lai, L. (2023). Guest Editorial: AI-Powered
Telco Network Automation: 5G Evolution and 6G. IEEE Wireless Communications, 30(1), 68–69.
[2] Militano, L., Zafeiropoulos, A., Fotopoulou, E., Bruschi, R., Lombardo, C., Edmonds, A., &
Papavassiliou, S. (2021, December). Ai-powered infrastructures for intelligence and automation in
beyond-5g systems. In 2021 IEEE Globecom Workshops (GC Wkshps) (pp. 1–6). New York City: IEEE.
[3] Asif, R., & Ghanem, K. (2021, January). Ai secured SD-WAN architecture as a latency critical IOT
enabler for 5G and beyond communications. In 2021 IEEE 18th Annual Consumer Communications &
Networking Conference (CCNC) (pp. 1–6). IEEE.
306 B. Girimurugan et al.
[4] Nguyen, H. X., Trestian, R., To, D., & Tatipamula, M. (2021). Digital twin for 5G and beyond. IEEE
Communications Magazine, 59(2), 10–15.
[5] Gallego-Madrid, J., Sanchez-Iborra, R., Ortiz, J., & Santa, J. (2023). The role of vehicular applications in
the design of future 6G infrastructures. ICT Express.
[6] Zeb, S., Mahmood, A., Khowaja, S. A., Dev, K., Hassan, S. A., Qureshi, N. M. F., . . . & Bellavista,
P. (2022). Industry 5.0 is coming: A survey on intelligent nextG wireless networks as technological
enablers. In 2022 IEEE International Conference on Networking and Internet Architecture (pp. 1–8).
[7] Singh, R., & Gill, S. S. (2023). Edge AI: A survey. Internet of Things and Cyber-Physical Systems, 3, 71–
92.
[8] Al-Mekhlal, M., AlYahya, A., Aldhamin, A., & Khan, A. (2022, August). Network automation python-
based application: The performance of a Multi-Layer Cloud Based Solution. In 2022 IEEE
International Conference on Omni-layer Intelligent Systems (COINS) (pp. 1–8). IEEE.
[9] Peltonen, E., Bennis, M., Capobianco, M., Debbah, M., Ding, A., Gil-Castiñeira, F., . . . & Yang,
T. (2020). 6G white paper on edge intelligence. In 2020 IEEE International Conference on
Distributed, Parallel, and Cluster Computing (pp. 1–6).
[10] Ordonez-Lucena, J., Ameigeiras, P., Contreras, L. M., Folgueira, J., & López, D. R. (2021). On the
rollout of network slicing in carrier networks: A technology radar. Sensors, 21(23), 8094.
[11] Pham, X. Q., Huynh-The, T., & Kim, D. S. (2022, October). A vision of distributed cloud computing. In
2022 13th International Conference on Information and Communication Technology Convergence
(ICTC) (pp. 1963–1968). IEEE.
[12] Ekudden, E. (2021). Five network trends: Towards the 6G Era. Ericsson Technology Review, 2021(9),
2–10.
[13] Thiruvasagam, P. K., Venkataram, V., Ilangovan, V. R., Perapalla, M., Payyanur, R., & Kumar,
V. (2023). Open RAN: Evolution of architecture, deployment aspects, and future directions. In 2023
IEEE International Conference on Networking and Internet Architecture (pp. 1–8).
[14] Pham, X. Q., Nguyen, T. D., Huynh-The, T., Huh, E. N., & Kim, D. S. (2022). Distributed cloud
computing: Architecture, enabling technologies, and open challenges. IEEE Consumer Electronics
Magazine, 12(3), 98–106.
[15] Wu, Y., Dai, H. N., Wang, H., Xiong, Z., & Guo, S. (2022). A survey of intelligent network slicing
management for industrial IoT: Integrated approaches for smart transportation, smart energy, and
smart factory. IEEE Communications Surveys and Tutorials, 24(2), 1175–1211.
[16] Singh, S., Karimipour, H., HaddadPajouh, H., & Dehghantanha, A. (2020). Artificial intelligence and
security of industrial control systems. In Handbook of Big Data Privacy, United States: Springer
International Publishing, 2020, pp. 121–164.
[17] Henttinen, T. (2021). AI SUPPORTED PURCHASE INVOICE POSTING: Lessons learned from piloting
project and future perspectives Master Thesis, Department of Business Management, Turku
University of Applied Science, pp.1–59.
[18] Firouzi, F., Jiang, S., Chakrabarty, K., Farahani, B., Daneshmand, M., Song, J., & Mankodiya, K. (2022).
Fusion of IoT, AI, Edge-Fog-Cloud, and Blockchain: Challenges, solutions, and a case study in
healthcare and medicine. IEEE Internet of Things Journal, 10(5), 3686–3705.
[19] Yadav, S. P., Agrawal, K. K., Bhati, B. S., Al-Turjman, F., & Mostarda, L. (2022). Blockchain-based
cryptocurrency regulation: An overview. Computational Economics, 59(4), 1659–1675.
[20] Irfan, D., Tang, X., Narayan, V., Mall, P. K., Srivastava, S., & Saravanan, V. (2022). Prediction of Quality
Food Sale in Mart Using the AI-Based TOR Method. Journal of Food Quality, 2022, 1–13.
[21] Sivasankari, S. S., Surendiran, J., Yuvaraj, N., Ramkumar, M., Ravi, C. N., & Vidhya, R. G. (2022, April).
Classification of diabetes using multilayer perceptron. In 2022 IEEE International Conference on
Distributed Computing and Electrical Circuits and Electronics (ICDCECE) (pp. 1–5). IEEE.
14 The smart and secured AI-powered strategies for optimizing processes 307
[22] Avinash Khatri, K. C., Shah, K. B., Logeshwaran, J., & Shrestha, A. (2023). Genetic algorithm based
techno-economic optimization of an isolated hybrid energy system. ICTACT Journal on
Microelectronics, 8(4), 1447–1450.
[23] Yadav, S. P., & Yadav, S. (2020). Fusion of medical images in wavelet domain: A hybrid
implementation. Computer Modeling in Engineering and Sciences, 122(1), 303–321.
[24] Rajalakshmi, M., Saravanan, V., Arunprasad, V., Romero, C. T., Khalaf, O. I., & Karthik, C. (2022).
Machine learning for modeling and control of industrial clarifier process. Intelligent Automation
and Soft Computing, 32(1), 339–359.
[25] Veerappan Kousik, N. G., Natarajan, Y., Suresh, K., Patan, R., & Gandomi, A. H. (2020). Improving power
and resource management in heterogeneous downlink OFDMA networks. Information, 11(4), 203.
[26] Ramesh, G., Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). Prediction of energy production level
in large PV plants through AUTO-Encoder Based Neural-Network (AUTO-NN) with restricted
Boltzmann feature extraction. Future Internet, 15(2), 46.
[27] Yadav, S. P., & Yadav, S. (2018). Fusion of medical images in wavelet domain: A Discrete
Mathematical Model/Fusion de imagenes medicas en el ambito de la transformada de ondicula:
modelo matematico discreto/Fusao de imagens medicas no ambito da transformada de wavelet:
Modelo matematico discreto. Revista Ingenieria Solidaria, 14(25), NA–NA.
[28] Saravanan, V., & Neeraja, A. (2013, January). Security issues in computer networks and
stegnography. In 2013 7th International Conference on Intelligent Systems and Control (ISCO)
(pp. 363–366). IEEE.
[29] Brilly Sangeetha, S., Wilfred Blessing, N. R., Yuvaraj, N., & Adeline Sneha, J. (2020). Improving the
training pattern in back-propagation neural networks using holt-winters’ seasonal method and
gradient boosting model. Applications of Machine Learning, 2020, 189–198.
[30] Adhikari, N., Logeshwaran, J., & Kiruthiga, T. (2022). The artificially intelligent switching framework
for terminal access provides smart routing in modern computer networks. BOHR International
Journal of Smart Computing and Information Technology, 3(1), 45–50.
[31] Pujahari, R. M., Yadav, S. P., & Khan, R. (2022). Intelligent farming system through weather forecast
support and crop production. In: Mohammad Ayoub Khan, Rijwan Khan and Mohammad Aslam
Ansari (Eds.). Application of Machine Learning in Agriculture (pp. 113–130). United States: Academic
Press.
[32] Khan, S., Saravanan, V. N. G. C., Lakshmi, T. J., Deb, N., & Othman, N. A. (2022). Privacy protection of
healthcare data over social networks using machine learning algorithms. Computational
Intelligence and Neuroscience, 2022, 1–12.
[33] Yuvaraj, N., Praghash, K., & Karthikeyan, T. (2022). Data privacy preservation and trade-off balance
between privacy and utility using deep adaptive clustering and elliptic curve digital signature
algorithm. Wireless Personal Communications, 124(1), 655–670.
[34] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. L.-R. U. B. I. (2023). An efficient load-based
resource utilization algorithm for bi-partite scatternet in wireless personal area networks.
International Journal of Communication Systems, 36(4), 1–17.
[35] Vermesan, O., Coppola, M., Nava, M. D., Capra, A., Kornaros, G., Bahr, R., . . . & McGrath, S. (2020).
New waves of IoT technologies research–transcending intelligence and senses at the edge to create
multi experience environments. Internet of Things–The Call of the Edge. Everything Intelligent
Everywhere, Europe: River Publisher, 1–168.
[36] Sandeepa, C., Siniarski, B., Kourtellis, N., Wang, S., & Liyanage, M. (2022). A survey on privacy for
B5G/6G: New privacy challenges, and research directions. Journal of Industrial Information
Integration, In 2022 IEEE International Conference on Cryptography and Security (pp. 1–49).
S. Rajeshwari, D. Praveenadevi, S. Revathy, S. P. Sreekala
15 Utilizing AI technologies to enhance
e-commerce business operations
Abstract: E-commerce has become an increasingly important sector for businesses,
offering customers convenient and fast access to goods and services. With the prolif-
eration of technological advancement, particularly Artificial intelligence (AI) technol-
ogy, e-commerce continues to develop and evolve. AI technologies such as machine
and deep learning can be utilized to enhance e-commerce operations. Companies can
employ AI-based solutions for personalized product recommendations, automated
customer service, optimized search and user interfaces, and improved security. By
utilizing AI technologies, e-commerce companies can improve customer experiences,
reduce operational costs, and make business operations more efficient.
https://doi.org/10.1515/9783111323749-015
310 S. Rajeshwari et al.
Potential challenges:
– Cost: AI can be expensive to implement and maintain. Companies need to invest
in hardware, software, and specialized teams to properly use and manage AI.
– Data quality: AI relies on data to produce accurate results. Poor quality data can
lead to inaccurate results, which can be costly for businesses.
15 Utilizing AI technologies to enhance e-commerce business operations 311
– Privacy and security: AI systems collect and process sensitive customer data, creat-
ing security risks and potential privacy violations. Companies must set policies and
procedures in place to protect customer data and ensure it is used responsibly.
– Association rules are a type of machine learning algorithm used to find relation-
ships between items in large datasets. This type of algorithm can be used for rec-
ommendation systems, market basket analysis, and other areas.
– Deep learning models are a type of machine learning algorithm used to model
complex relationships between input and output data. Deep learning models are
used in a wide variety of applications, such as object detection, image recognition,
natural language processing (NLP), and recommendation systems. These models
can analyze large datasets and learn patterns and trends that are difficult for hu-
mans to identify.
Together, intelligent search and NLP can enable computers to process natural lan-
guage in a more meaningful way and generate more intelligent search results. This
can help to provide users with more accurate and relevant information, as well as
create an improved user experience for interface designers.
AI-driven search engines are search engines that use AI to provide more accurate and
relevant search results. AI-driven search engines can incorporate machine learning
algorithms, NLP algorithms, and use deep learning techniques to understand user in-
tent better and return more comprehensive results for a given query [6]. These tech-
niques are used to improve the accuracy of the search results and minimize the
amount of time required to sift through irrelevant information. Additionally, AI-
driven search engines are leveraging intelligent search algorithms to return the most
up-to-date results so that users always have the most current information. Intelligent
search algorithms dynamically adjust search parameters based on the latest data
from a variety of sources in order to return the most relevant set of results. This re-
sults in more targeted and relevant search results and enables users to find the most
accurate and up-to-date answers to their search queries.
Voice search and chatbot applications are becoming increasingly important for
e-commerce businesses as they allow customers to quickly access information and easily
make transactions. Voice search allows customers to search for products and services in
natural language, making it easier and faster for customers to find what they are looking
for. Chatbot applications provide customers with an interactive experience that is per-
sonalized to their needs and can help to provide quick answers to customer queries.
Chatbots can also help to provide customers with promotional offers, further increasing
customer engagement [8]. These applications provide a more immersive and personal-
ized shopping experience for customers and can lead to customer loyalty and increased
sales. In addition, these applications can help businesses to reduce costs associated with
providing customer service, as they can take over many of the duties performed by
employees.
Behavioral analytics is the practice of using, collecting and analyzing data pertaining
to user behavior to better understand and inform decisions about product, marketing,
and other aspects of modern businesses. Anomaly detection techniques are used to
detect suspicious behavior or patterns that don’t conform to expected norms. They
can be used to detect fraud, malware, and other kinds of malicious activities [16].
Anomaly detection techniques typically involve applying machine learning algorithms
to identify suspicious or abnormal patterns. These techniques leverage data mining,
statistical methods, text analytics, and other techniques to analyze and identify un-
usual or outlier events in a dataset.
AI-driven demand forecasting and inventory optimization is a process that helps busi-
nesses to accurately predict customer demand for products and services. It uses AI to
determine the best inventory levels to optimize profits while minimizing waste and
inefficiencies. AI algorithms collect data from sources such as seasonality, customer
feedback, past order rates, environmental factors, and economic changes to create
predictions. These predictions are then used to calculate the optimal inventory levels
for the company to keep on hand [21]. AI-driven demand forecasting and inventory
optimization enable businesses to reduce the cost of indexing, increase customer sat-
isfaction through better service, and improve stock management.
15 Utilizing AI technologies to enhance e-commerce business operations 319
ces and maximize revenue. This technology offers an efficient and accurate solution to
the issue of manual price-setting, which can be labor-intensive and prone to errors. AI-
driven dynamic pricing algorithms can quickly respond to fluctuations in demand and
competitor prices, facilitated by the use of machine learning models that continuously
learn and optimize in real-time. The result is optimal pricing customized to the custom-
er’s needs and continuously adjusted to suit changing market conditions.
AI algorithms for dynamic pricing are designed to optimize the pricing of products
and services according to current market conditions. The algorithms look at factors
like supply and demand, perceived value, and competitor pricing in order to optimize
a product’s pricing. These algorithms enable companies to set and adjust prices in
real time, automatically responding to changes in the market. Additionally, AI algo-
rithms can help businesses analyze competitor pricing strategies and identify oppor-
tunities to set competitive prices [24]. By taking into account market trends, product
availability, and an understanding of consumer behavior, these algorithms can pro-
vide businesses with valuable insights on how to set competitive prices. Ultimately,
using AI algorithms for dynamic pricing and competitive pricing analysis can help
businesses make better pricing decisions and maximize their profits.
and profitable. It can also be used to create highly customized and individualized pric-
ing models for each customer. Additionally, price optimization techniques are becoming
increasingly important for businesses in volatile markets, as they allow companies to
quickly adjust prices in response to market trends.
Challenges:
– AI requires a considerable amount of data in order to produce effective insights.
Without adequate data, AI solutions may be unable to optimize pricing strategies.
– Additionally, AI solutions are not foolproof and may produce inaccurate results.
Businesses need to regularly audit and adjust their pricing strategies in order to
ensure accuracy.
322 S. Rajeshwari et al.
NLP is the technology and discipline of extracting meaningful information from natu-
ral language text. NLP can help automate customer interactions by allowing com-
puters to understand natural language. Sentiment analysis is a type of NLP technique
that uses algorithms to interpret a customer’s text to understand the sentiment the
language conveys. It can be used to identify customer sentiment in emails, online re-
views, surveys, social media posts, and other customer interactions.
15 Utilizing AI technologies to enhance e-commerce business operations 323
being used, as well as to ensure that sensitive data is protected and collected ethically.
Algorithmic fairness is also essential in order to avoid bias or discrimination, such as
intentional or unintentional errors and inaccuracies in functional AI tasks. By adhering
to these three factors, companies can create more transparent and fair AI solutions.
Privacy and data protection considerations in e-commerce refer to the steps taken to
ensure that customer data is kept secure and protected from potential abuse or mis-
use. This includes measures such as encryption, strong authentication, and secure
transmission protocols. Additionally, when collecting personal data, businesses should
make sure to collect only the data necessary for their operations, and make sure cus-
tomers are made aware of how their data is used or shared. Furthermore, businesses
should have processes in place for overseeing how collected data is used, stored, and
secured. Ultimately, businesses should take all reasonable steps to ensure they are
compliant with applicable data protection regulation.
accountability regarding how decisions are made. Responsible AI also entails taking
steps to meet privacy standards and ensure data protection for consumers and the
business.
Mitigating bias in e-commerce operations start with understanding how personal
data is collected and used by e-commerce platforms. It’s important to also understand
the algorithms used to make recommendations and decisions. Companies should
strive to create properly functioning and accountable AI systems that make decisions
in a way that takes customer needs and preferences into account [34]. To do this, it is
necessary to establish AI models that are regularly checked and maintained to ensure
accuracy and consistency with set rules. Companies should also monitor AI models to
identify any areas of bias and take corrective action. Finally, companies should con-
sider diversity among their data scientists and other team members to enable a range
of unique perspectives to guide the development and implementation of AI models.
likely that AI technology in e-commerce will become even smarter and more integrated
into the user experience. As machine learning and NLP become more pervasive, more
platforms and applications will be able to understand customer requests and provide
them with personalized and automated assistance. This will result in improved cus-
tomer experience, greater customer loyalty, and increased sales.
AI-driven personalization and hyper-targeted marketing are two different tools used
to enhance marketing campaigns. AI-driven personalization is the use of advanced al-
gorithms to create personalized content and experiences for customers that are tai-
lored to each individual. This involves using AI to automatically analyze customer
data, make predictions about what the customer wants, and then provide content that
is tailored to the customer’s preferences [36].
Hyper-targeted marketing is the use of customer data and other data sources to
target an audience with a specific message or offer. This involves segmenting the au-
dience into groups based on their characteristics or interests and then creating mar-
keting messages tailored to each individual group. By using hyper-targeted marketing,
companies can ensure that their message is being delivered to the right people.
328 S. Rajeshwari et al.
References
[1] Song, X., Yang, S., Huang, Z., & Huang, T. (2019, August). The application of artificial intelligence in
electronic commerce. Journal of Physics: Conference Series, 1302(3), 032030. IOP Publishing.
[2] Khrais, L. T. (2020). Role of artificial intelligence in shaping consumer demand in E-commerce.
Future Internet, 12(12), 226.
[3] Xue, M., Xiu, G., Saravanan, V., & Montenegro-Marin, C. E. (2021). Cloud computing with AI for
banking and e-commerce applications. The Electronic Library, 39(4), 539–552.
[4] Bawack, R. E., Wamba, S. F., Carillo, K. D. A., & Akter, S. (2022). Artificial intelligence in E-Commerce:
A bibliometric study and literature review. Electronic Markets, 32(1), 297–338.
[5] Zhang, D., Pee, L. G., & Cui, L. (2021). Artificial intelligence in E-commerce fulfillment: A case study of
resource orchestration at Alibaba’s Smart Warehouse. International Journal of Information
Management, 57, 102304.
[6] Yang, X. (2019, June). Satisfaction evaluation and optimization of tourism e-commerce users based
on artificial intelligence technology. In 2019 International Conference on Robots & Intelligent
System (ICRIS) (pp. 373–375). IEEE.
15 Utilizing AI technologies to enhance e-commerce business operations 329
[7] George, A. S., & George, A. H. (2023). A review of ChatGPT AI’s impact on several business sectors.
Partners Universal International Innovation Journal, 1(1), 9–23.
[8] Ikumoro, A. O., & Jawad, M. S. (2019). Intention to use intelligent conversational agents in
e-commerce among Malaysian SMEs: an integrated conceptual framework based on tri-theories
including unified theory of acceptance, use of technology (UTAUT), and TOE. International Journal of
Academic Research in Business and Social Sciences, 9(11), 205–235.
[9] Toleuuly, A., Yessengeldin, B., Jumabaeva, S., & Zhanseitov, A. (2019). Contemporary problems and
prospects of e-commerce development in modern conditions. Revista Espacios, 40(06), 1–6.
[10] Mnyakin, M. (2020). Investigating the impacts of AR, AI, and Website Optimization on Ecommerce
Sales Growth. ResearchBerg Review of Science and Technology, 3(1), 116–130.
[11] Micu, A., Micu, A. E., Geru, M., Căpățînă, A., & Muntean, M. C. (2021). The Impact of artificial
intelligence use on E-Commerce in Romania. Amfiteatru Economic, 23(56), 137–154.
[12] Pallathadka, H., Ramirez-Asis, E. H., Loli-Poma, T. P., Kaliyaperumal, K., Ventayen, R. J. M., & Naved,
M. (2023). Applications of artificial intelligence in business management, e-commerce and finance.
Materials Today: Proceedings, 80, 2610–2613.
[13] Li, S. (2020). Structure optimization of e-commerce platform based on artificial intelligence and
blockchain technology. Wireless Communications and Mobile Computing, 2020, 1–8.
[14] Maslak, O. I., Maslak, M. V., Grishko, N. Y., Hlazunova, O. O., Pererva, P. G., & Yakovenko, Y. Y. (2021,
September). Artificial intelligence as a key driver of business operations transformation in the
conditions of the digital economy. In 2021 IEEE International Conference on Modern Electrical and
Energy Systems (MEES) (pp. 1–5). IEEE.
[15] Panigrahi, D., & Karuna, M. (2021). A review on leveraging artificial intelligence to enhance business
engagement in ecommerce. Journal Homepage: www.ijrpr.com ISSN, 2582(7421), 2.
[16] Moriuchi, E., Landers, V. M., Colton, D., & Hair, N. (2021). Engagement with chatbots versus
augmented reality interactive technology in e-commerce. Journal of Strategic Marketing, 29(5),
375–389.
[17] Rani, P., Verma, S., Yadav, S. P., Rai, B. K., Naruka, M. S., & Kumar, D. (2022). Simulation of the
lightweight blockchain technique based on privacy and security for healthcare data for the cloud
system. International Journal of E-Health and Medical Communications (IJEHMC), 13(4), 1–15.
[18] Khan, S., Saravanan, V. N. G. C., Lakshmi, T. J., Deb, N., & Othman, N. A. (2022). Privacy protection of
healthcare data over social networks using machine learning algorithms. Computational
Intelligence and Neuroscience, 2022, 1–9.
[19] Yuvaraj, N., & Vivekanandan, P. (2013, February). An efficient SVM based tumor classification with
symmetry non–negative matrix factorization using gene expression data. In 2013 International
Conference on Information Communication and Embedded Systems (Icices) (pp. 761–768). New
York: IEEE.
[20] Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model
for efficient resource allocation in 5G broadband communication networks. ICTACT Journal on Soft
Computing, 13(3), 2986–2994.
[21] Yadav, S. P., Zaidi, S., Mishra, A., & Yadav, V. (2022). Survey on machine learning in speech emotion
recognition and vision systems using a recurrent neural network (RNN). Archives of Computational
Methods in Engineering, 29(3), 1753–1770.
[22] Saravanan, V., & Mohan Raj, V. (2016). Maximizing QoS by cooperative vertical and horizontal
handoff for tightly coupled WiMAX/WLAN overlay networks. Cluster Computing, 19(3), 1619–1633.
[23] Yuvaraj, N., Chang, V., Gobinathan, B., Pinagapani, A., Kannan, S., Dhiman, G., & Rajan, A. R. (2021).
Automatic detection of cyberbullying using multi-feature based artificial intelligence with deep
decision tree classification. Computers and Electrical Engineering, 92, 107186.
330 S. Rajeshwari et al.
[24] Ramesh, G., Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). Prediction of energy production level
in large PV plants through AUTO-Encoder Based Neural-Network (AUTO–NN) with restricted
Boltzmann feature extraction. Future Internet, 15(2), 46.
[25] Yadav, S. P., Mahato, D. P., & Linh, N. T. D. (Eds.). (2020). Distributed artificial intelligence: A modern
approach. Boca Raton: CRC Press.
[26] Munagala, N. V. L., Saravanan, V., Almukhtar, F. H., Jhamat, N., Kafi, N., & Khan, S. (2022). Supervised
approach to identify autism spectrum neurological disorder via label distribution learning.
Computational Intelligence and Neuroscience, 2022, 1–11.
[27] Gobinathan, B., Mukunthan, M. A., Surendran, S., Somasundaram, K., Moeed, S. A., Niranjan, P., . . .
& Sundramurthy, V. P. (2021). A novel method to solve real time security issues in software industry
using advanced cryptographic techniques. Scientific Programming, 2021, 1–9.
[28] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). L-RUBI: An efficient load-based
resource utilization algorithm for bi-partite scatternet in wireless personal area networks.
International Journal of Communication Systems, 36(4), 1–17.
[29] Yadav, S. P., & Yadav, S. (2020). Fusion of medical images in wavelet domain: A hybrid
implementation. Computer Modeling in Engineering and Sciences, 122(1), 303–321.
[30] Saravanan, V., & Sumathi, A. (2012, December). Dynamic handoff decision based on current traffic
level and neighbor information in wireless data networks. In 2012 Fourth International Conference
on Advanced Computing (ICoAC) (pp. 1–5). IEEE.
[31] Raja, S., Logeshwaran, J., Venkatasubramanian, S., Jayalakshmi, M., Rajeswari, N., Olaiya, N. G., &
Mammo, W. D. (2022). OCHSA: Designing energy-efficient lifetime-aware leisure degree adaptive
routing protocol with optimal cluster head selection for 5G communication network disaster
management. Scientific Programming, 2022, 1–12.
[32] Sivaram, M., Yuvaraj, D., Mohammed, A. S., Manikandan, V., Porkodi, V., & Yuvaraj, N. (2019).
Improved enhanced DBTMA with contention-aware admission control to improve the network
performance in MANETS. CMC-Computers Materials and Continua, 60(2), 435–454.
[33] Sandu, N., & Gide, E. (2019, September). Adoption of AI–Chatbots to enhance student learning
experience in higher education in India. In 2019 18th International Conference on Information
Technology Based Higher Education and Training (ITHET) (pp. 1–5). IEEE.
[34] Ping, N. L. (2019, December). Constructs for artificial intelligence customer service in E-commerce.
In 2019 6th International Conference on Research and Innovation in Information Systems (ICRIIS)
(pp. 1–6). IEEE.
[35] Agnihotri, R. (2021). From sales force automation to digital transformation: How social media, social
CRM, and artificial intelligence technologies are influencing the sales process. A Research Agenda
for Sales, 2021, 21–47.
[36] Zoroja, J., Klopotan, I., & Ana-Marija, S. (2020). Quality of e-commerce practices in European
enterprises: Cluster analysis approach. Interdisciplinary Description of Complex Systems: INDECS,
18(2–B), 312–326.
Vijaya Gunturu, Charanjeet Singh, Nikhil S. Patankar, S. Praveena
16 Exploring the potential of artificial
intelligence in wireless sensor networks
Abstract: In various monitoring, tracking, and control applications, wireless sensor
networks (WSNs) are gaining popularity. To fully utilize the promise of WSNs in vari-
ous applications, two significant challenges – power efficiency and scalability – must
be overcome. Recent developments in artificial intelligence (AI) methods, such as
deep learning, machine learning (ML), and reinforcement learning, present fresh per-
spectives and development opportunities for WSNs to make wise decisions and effec-
tively use resources. For example, AI-based techniques can enable WSNs to learn the
underlying patterns and trends in the data streams exchanged among different sensor
nodes, and choose optimal parameters for efficient data collection and analysis.
16.1 Introduction
The utilization of AI techniques can lead to smart, adaptive strategies in WSNs which
are capable of self-organization and to quickly react to external dynamic environ-
ments. Therefore, AI-based methods could greatly improve the power efficiency, scal-
ability, flexibility, and robustness of WSNs [1]. This paper provides an overview of
recent advancements in AI techniques and explores potential applications in WSNs.
Furthermore, potential research opportunities are discussed, and open challenges are
identified in order to further leverage the potential of AI for WSNs.
https://doi.org/10.1515/9783111323749-016
332 Vijaya Gunturu et al.
enabling efficient resource consumption [3]. Furthermore, AI can ease the stress on a
network’s communication by enabling automated decision-making.
The WSNs are networks of interconnected nodes that allow for data collection
and processing with minimal dependence on wired connections. With their growing
popularity and affordability, they are becoming more common for applications from
industrial sensing to IoT applications [9].
Challenges in WSN:
– Due to its reliance on radio-based communication that can be intercepted, secu-
rity continues to be one of the largest difficulties facing WSN. To keep the net-
work and its data safe from hostile actions, security needs to be addressed.
– Network management and maintenance: Given the abundance of sensor nodes
inside a WSN and the limited resources available, network management and
maintenance can frequently be a challenging undertaking.
– Power management: Because WSN sensor nodes have a certain amount of power
available to them, careful power management is necessary to guarantee the net-
work functions properly.
– Data collection: Since real-time data is lacking in WSNs, gathering data might be
challenging.
– Scalability: The WSN needs to be scalable in order to quickly fulfill the need for
increased data and resources.
Algorithms that employ statistical methods to help machines learn from data without
explicit programming are known as machine learning algorithms. They are employed
to base predictions or choices on information. These techniques include support vec-
tor machines, decision trees, random forests, logistic regression, and linear regression
as examples. Numerous industries, including finance, healthcare, marketing, and
computer vision, use these algorithms. They can also be utilized for tasks like fraud
detection, stock market analysis, facial recognition, and text classification [18].
One form of machine learning algorithm that extracts knowledge from labeled data is
supervised learning. The objective of supervised learning is to create a model that can
predict outcomes based on fresh data. Predictive analysis and categorization issues
are typically addressed using it. In supervised learning, the algorithm is trained on
labeled data in order to produce a function that can be used to predict outcomes from
fresh data [19].
336 Vijaya Gunturu et al.
Unsupervised learning is a kind of machine learning technique that uses data analysis
without the need for prior labeling of the data. Without receiving any outside direc-
tion, it tries to find data patterns and features that are connected to or resemblant to
one another. As a result, it’s frequently used to find clusters in huge datasets and to
uncover information that might not have been discovered otherwise. Algorithms for
unsupervised learning can be utilized for a variety of applications, including anomaly
detection, dimensionality reduction, and grouping [20].
By receiving feedback in the form of incentives or penalties, an agent can learn be-
haviors from its environment using a machine learning technique called reinforce-
ment learning. It is based on the concept of learning from mistakes, and it provides
the agent with the opportunity to do so by exploring and exploiting new opportuni-
ties. The agent will learn behaviors to maximize its expected reward for a given task
by making observations and taking actions as it learns how to interact with its envi-
ronment. Reinforcement learning is used in many applications, such as robotics, self-
driving cars, finance, and control systems [21].
A type of AI called deep learning uses numerous layers of neurons to process input
and produce patterns. Deep learning algorithms can forecast the data and discover
brand-new patterns that were previously undetected by using these patterns. Health-
care, finance, marketing, and other industries can all benefit from the use of this tech-
nology. Deep learning’s capacity to find latent properties in data is its key benefit. It
can identify data patterns that are too complicated for humans to notice [22]. Because
it can swiftly and reliably process massive amounts of data, this technology is growing
in popularity. Due to this, many organizations are now utilizing deep learning to
make decisions and boost their production and efficiency.
Swarm intelligence algorithms are computer algorithms designed to mimic the behav-
ior of swarms or groups of organisms. These algorithms are used mainly in robotics
and AI, and they are inspired by how real-world swarms usually act. They simulate col-
lective behavior and the ability of swarms to learn from their environment. Swarm in-
16 Exploring the potential of artificial intelligence in wireless sensor networks 337
telligence algorithms are used to solve complex problems or make decisions where no
single entity has control [23]. These algorithms rely on agents working cooperatively
and exchanging information so they can learn from each other.
Data fusion and information processing in AI-based WSN are both key elements of
managing and analyzing data that is collected by WSNs. The process of merging data
338 Vijaya Gunturu et al.
The process of locating sensor nodes (or devices) physically in WSNs involves exploiting
the infrastructure that is already in place, such as beacons, GPS, accelerometers, or
other wireless communication systems. Numerous WSN applications, such as surveil-
lance, asset tracking, and environmental monitoring, depend on localization. As part of
a tracking and localization loop, tracking allows the network to keep track of how the
device is moving around the environment and to update the device’s location [28].
Ranging methods (triangulation, trilateration, etc.) and positioning methods are
two groups of localization algorithms. Both approaches rely on the ability to deter-
ministically measure distances or angles to beacons (e.g., Wi-Fi APs and stationary
nodes) or other peers in the network. Ranging methods, also referred to as view-
based approaches, are typically based on measuring received signal strength or time
of arrival. Positioning methods are based on GPS or UWB which provide highly pre-
cise coordinates for sensor nodes [29].
Tracking in WSNs can be divided into two categories as well: motion tracking and
tracking with respect to absolute location. Motion tracking involves tracking node
movements within a given area (e.g., indoor navigation) and is usually combined with
position-based localization algorithms. Tracking with respect to absolute location can
be used to determine the position of sensor nodes in a larger scale environment (e.g.,
indoor or outdoor) and requires ranging or positioning methods [30].
Localization and tracking technologies form the backbone of many WSN applica-
tions such as robotics, environment monitoring, or asset tracking. By providing accu-
rate location and positioning information, they enable applications to react and adapt
to the environment surrounding them.
16 Exploring the potential of artificial intelligence in wireless sensor networks 339
A method for detecting and diagnosing flaws in WSNs is known as fault detection and
diagnosis in AI in WSNs. This method uses machine learning and AI algorithms to
find faults. In order to detect errors and determine their root cause, the AI algorithms
used for fault detection examine network data and extract information from it. The
AI algorithms then provide a diagnosis to the user about the fault’s origin and the
best course of action to remedy it [31]. This process is useful for autonomous systems,
such as for example unmanned aerial vehicles (UAVs), as it allows them to self-
diagnose and self-heal in case of a failure. AI-based fault detection and diagnosis can
be used in a variety of settings, including medical, industrial, and military systems.
Intrusion detection and prevention is a technology designed to identify and prevent ma-
licious activity in a computer network. It is a combination of software and hardware
products that detect, monitor, and alert system administrators to suspicious activity. In-
trusion detection and prevention systems (IDPS) are used to detect unauthorized access
to networked systems and applications. They can also provide protection against mal-
ware, unauthorized access to system resources, data theft, and denial of service attacks
[33]. IDPS can be used to identify malicious activity by monitoring network traffic for
signs of malicious behavior and alerting administrators to potential threats. Addition-
ally, an IDPS can be configured to take action to block or deny access to malicious traf-
fic when it is detected.
340 Vijaya Gunturu et al.
Secure data transmission is an approach for securely exchanging data over a network.
In secure data transmission, endpoints encrypt communications with the use of as-
ymptotically secure encryption algorithms such as TLS, IPSec, and SSH. The encryp-
tion and authentication can be done for the entire data packet or only the meta-data.
Additional security protocols such as PKI (public key infrastructure) and authentica-
tion methods like Kerberos are used in order to ensure that data is transferred with-
out being tampered by malicious actors. It also ensures that the data is transferred
only to authorized users and that the communication is only between the authorized
parties. Secure data transmission is essential for sensitive data such as financial data,
medical records, and other confidential information [34].
References
[1] Ghazal, T. M. (2022). Data fusion-based machine learning architecture for intrusion detection.
Computers, Materials and Continua, 70(2), 3399–3413.
[2] Howard, J. (2019). Artificial intelligence: Implications for the future of work. American Journal of
Industrial Medicine, 62(11), 917–926.
[3] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). L-RUBI: An efficient load-based
resource utilization algorithm for bi-partite scatternet in wireless personal area networks.
International Journal of Communication Systems, 36(4), 1–17.
[4] Lăzăroiu, G., Kliestik, T., & Novak, A. (2021). Internet of things smart devices, industrial artificial
intelligence, and real-time sensor networks in sustainable cyber-physical production systems.
Journal of Self-Governance and Management Economics, 9(1), 20–30.
[5] Mekonnen, Y., Namuduri, S., Burton, L., Sarwat, A., & Bhansali, S. (2019). Machine learning
techniques in wireless sensor network based precision agriculture. Journal of the Electrochemical
Society, 167(3), 037522.
[6] Ali, S. S., & Choi, B. J. (2020). State-of-the-art artificial intelligence techniques for distributed smart
grids: A review. Electronics, 9(6), 1030.
[7] Sawhney, R., Malik, A., Sharma, S., & Narayan, V. (2023). A comparative assessment of artificial
intelligence models used for early prediction and evaluation of chronic kidney disease. Decision
Analytics Journal, 6, 100169.
[8] Adeel, A., Gogate, M., Farooq, S., Ieracitano, C., Dashtipour, K., Larijani, H., & Hussain, A. (2019). A
survey on the role of wireless sensor networks and IoT in disaster management. Geological Disaster
Monitoring Based on Sensor Networks, 10, 57–66.
[9] Nica, E., & Stehel, V. (2021). Internet of things sensing networks, artificial intelligence-based
decision-making algorithms, and real-time process monitoring in sustainable industry 4.0. Journal
of Self-Governance and Management Economics, 9(3), 35–47.
[10] Thangaramya, K., Kulothungan, K., Logambigai, R., Selvi, M., Ganapathy, S., & Kannan, A. (2019).
Energy aware cluster and neuro-fuzzy based routing algorithm for wireless sensor networks in IoT.
Computer Networks, 151, 211–223.
[11] Arroyo, P., Herrero, J. L., Suárez, J. I., & Lozano, J. (2019). Wireless sensor network combined with
cloud computing for air quality monitoring. Sensors, 19(3), 691.
[12] Yuvaraj, N., & Vivekanandan, P. (2013, February). An efficient SVM based tumor classification with
symmetry non-negative matrix factorization using gene expression data. In 2013 International
Conference on Information Communication and Embedded Systems (Icices) (pp. 761–768). IEEE.
[13] Yadav, S. P., & Yadav, S. (2019). Mathematical implementation of fusion of medical images in
continuous wavelet domain. Journal of Advanced Research in Dynamical and Control System, 10(10),
45–54.
[14] Yuvaraj, N., Chang, V., Gobinathan, B., Pinagapani, A., Kannan, S., Dhiman, G., & Rajan, A. R. (2021).
Automatic detection of cyberbullying using multi-feature based artificial intelligence with deep
decision tree classification. Computers and Electrical Engineering, 92, 107186.
[15] Vaniprabha, A., Logeshwaran, J., Kiruthiga, T., & Shah, K. B. (2022). Examination of the effects of
long-term COVID-19 impacts on patients with neurological disabilities using a neuro machine
learning model. BOHR International Journal of Neurology and Neuroscience, 1(1), 21–28.
[16] Saravanan, V., & Sumathi, A. (2012). Handoff mobiles with low latency in heterogeneous networks
for seamless mobility: A survey and future directions. European Journal of Scientific Research, 81(3),
417–424.
342 Vijaya Gunturu et al.
[17] Pujahari, R. M., Yadav, S. P., & Khan, R. (2022). Intelligent farming system through weather forecast
support and crop production. In: Mohammad Ayoub Khan, Rijwan Khan and Mohammad Aslam Ansari
(Eds.), Application of Machine Learning in Agriculture (pp. 113–130). United States: Academic Press.
[18] Mojjada, R. K., Yadav, A., Prabhu, A. V., & Natarajan, Y. (2020). Machine learning models for covid-19
future forecasting. Materials Today: Proceedings, 2020, 1–5.
[19] Yadav, S. P., & Yadav, S. (2018). Fusion of medical images in wavelet domain: A discrete
mathematical model. Ingeniería Solidaria, 14(25), 1–11.
[20] Tripathy, R., Nayak, R. K., Saravanan, V., Mishra, D., Parasa, G., Das, K., & Das, P. (2021). Spectral
clustering based fuzzy C-means algorithm for prediction of membrane cholesterol from ATP-binding
cassette transporters. In Intelligent and Cloud Computing: Proceedings of ICICC 2019 (Vol. 2,
pp. 439–448). Springer Singapore.
[21] Rani, P., Verma, S., Yadav, S. P., Rai, B. K., Naruka, M. S., & Kumar, D. (2022). Simulation of the
lightweight blockchain technique based on privacy and security for healthcare data for the cloud
system. International Journal of E-Health and Medical Communications (IJEHMC), 13(4), 1–15.
[22] Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J., . . . &
Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey
wolf optimization in wireless sensor networks. Security and Communication Networks, 2022, 1–13.
[23] Gopi, B., Logeshwaran, J., & Kiruthiga, T. (2022). An innovation in the development of a mobile radio
model for a dual-band transceiver in wireless cellular communication. BOHR International Journal of
Computational Intelligence and Communication Network, 1(1), 20–25.
[24] Saravanan, V., & Sumathi, A. (2012, December). Dynamic handoff decision based on current traffic
level and neighbor information in wireless data networks. In 2012 Fourth International Conference
on Advanced Computing (ICoAC) (pp. 1–5). IEEE.
[25] Gobinathan, B., Mukunthan, M. A., Surendran, S., Somasundaram, K., Moeed, S. A., Niranjan, P., . . .
& Sundramurthy, V. P. (2021). A novel method to solve real time security issues in software industry
using advanced cryptographic techniques. Scientific Programming, 2021, 1–9.
[26] Kakani, V., Nguyen, V. H., Kumar, B. P., Kim, H., & Pasupuleti, V. R. (2020). A critical review on
computer vision and artificial intelligence in food industry. Journal of Agriculture and Food
Research, 2, 100033.
[27] Wang, C. X., Di Renzo, M., Stanczak, S., Wang, S., & Larsson, E. G. (2020). Artificial intelligence
enabled wireless networking for 5G and beyond: Recent advances and future challenges. IEEE
Wireless Communications, 27(1), 16–23.
[28] Adhikari, N., Logeshwaran, J., & Kiruthiga, T. (2022). The artificially intelligent switching framework
for terminal access provides smart routing in modern computer networks. BOHR International
Journal of Smart Computing and Information Technology, 3(1), 45–50.
[29] Savaglio, C., Pace, P., Aloi, G., Liotta, A., & Fortino, G. (2019). Lightweight reinforcement learning for
energy efficient communications in wireless sensor networks. IEEE Access, 7, 29355–29364.
[30] Otoum, S., Kantarci, B., & Mouftah, H. T. (2019). On the feasibility of deep learning in sensor network
intrusion detection. IEEE Networking Letters, 1(2), 68–71.
[31] Gruson, D., Helleputte, T., Rousseau, P., & Gruson, D. (2019). Data science, artificial intelligence, and
machine learning: Opportunities for laboratory medicine and the value of positive regulation.
Clinical Biochemistry, 69, 1–7.
[32] Lussier, F., Thibault, V., Charron, B., Wallace, G. Q., & Masson, J. F. (2020). Deep learning and
artificial intelligence methods for Raman and surface-enhanced Raman scattering. TRAC: Trends in
Analytical Chemistry, 124, 115796.
[33] Nallaperuma, D., Nawaratne, R., Bandaragoda, T., Adikari, A., Nguyen, S., Kempitiya, T., . . . &
Pothuhera, D. (2019). Online incremental machine learning platform for big data-driven smart traffic
management. IEEE Transactions on Intelligent Transportation Systems, 20(12), 4679–4690.
16 Exploring the potential of artificial intelligence in wireless sensor networks 343
[34] Ghazal, T. M., Hasan, M. K., Alshurideh, M. T., Alzoubi, H. M., Ahmad, M., Akbar, S. S., . . . & Akour,
I. A. (2021). IoT for smart cities: Machine learning approaches in smart healthcare – a review. Future
Internet, 13(8), 218.
[35] Pandian, A. P. (2019). Artificial intelligence application in smart warehousing environment for
automated logistics. Journal of Artificial Intelligence, 1(02), 63–72.
[36] Kumar, D. P., Amgoth, T., & Annavarapu, C. S. R. (2019). Machine learning algorithms for wireless
sensor networks: A survey. Information Fusion, 49, 1–25.
M. V. Jagannatha Reddy, J. Somasekar, Kaushalya Thopate,
Sanjeev Kumar Angadi
17 Exploring artificial intelligence techniques
for enhanced sentiment analysis through
data mining
Abstract: Sentiment analysis is a subset of data mining that is used to locate, collect,
and examine people’s opinions. It makes it possible for data mining tools to compre-
hend peoples’ feelings and ideas more fully. Natural language processing is one of the
most used and established methods used for sentiment analysis tasks. This technique
is mainly based on rules, principles, and algorithms which aid in understanding the
text making it easier to segment and extract meaningful data. However, this tech-
nique is becoming obsolete due to the sheer volume and complexity of data present
these days.
17.1 Introduction
A new generation of smart algorithms such as artificial intelligence (AI) have been
developed for the improved sentiment analysis through data mining. AI techniques
such as deep learning networks, logic programming, and machine learning techni-
ques are proving to be two times more accurate than traditional sentiment analysis.
These AI techniques provide greater accuracy, flexibility, and efficiency in sentiment
analysis tasks [1]. They understand not only the sentiment of the data but also the con-
text in which the data is used. This helps in making the sentiment analysis more pre-
cise as it can also detect the intensity of sentiment expressed in the data. Therefore,
AI techniques can be used as an enhanced approach for sentiment analysis through
data mining.
The scope and objectives of the AI techniques for enhanced sentiment analysis through
data mining are to utilize intelligent technologies to automate and improve the accuracy
of sentiment analysis. Natural language processing (NLP) and supervised learning are
examples of AI approaches [2]. The AI-driven systems will be able to better identify the
latent associations within the data and, based on this, ascertain the sentiment of the text
by using data mining techniques like clustering, association rule mining, or sentiment
https://doi.org/10.1515/9783111323749-017
346 M. V. Jagannatha Reddy et al.
analysis. This might potentially be applied to supervised learning models, which train
models using labeled data. Moreover, the AI-driven system could also be used to identify
new themes that have not yet been identified by traditional sentiment analysis tools.
This method aims to offer sentiment analysis results that are more accurate than
those from conventional techniques [3]. The AI-driven approach should give a better
understanding of the text and the sentiment behind it, making the sentiment results
more accurate and specific. This is beneficial for organizations looking to gain more
insight into their users’ sentiments about their product or services.
customer feedback, companies can identify potential issues that they may not have
previously been aware of and address them before they become serious problems.
Additionally, sentiment analysis can be used to develop a better understanding of cus-
tomer preferences and develop more effective products and services. Ultimately, sen-
timent analysis helps companies to create a more positive customer experience.
The study of sentiment conveyed in text is the focus of the field of NLP known as sen-
timent analysis. The difficulty of the issue stems from the intricacy of language. Chal-
lenges in sentiment analysis include:
– Interpretation ambiguities: Natural language contains ambiguity. This can lead to
difficulty in interpreting the sentiment of a particular phrase. For example, the
phrase “I’m not bad” may be interpreted as having a positive sentiment or a nega-
tive sentiment, depending on the context.
– Context dependency: The sentiment of a phrase may depend on the context. For
example, the phrase “It’s going to be a great day” may be interpreted as having a
positive sentiment if it is uttered in the morning before the individual has en-
countered any difficulties, but may be interpreted as a negative sentiment if it is
uttered late in the evening after the individual has experienced setbacks.
– Temporal dependencies: The sentiment of a phrase may change over time. For
example, a phrase uttered during a historical moment or event can carry a differ-
ent sentiment than a phrase uttered after the event, when the current sentiment
surrounding it may have changed.
– Lexical ambiguities: The sentiment of a phrase may depend on the word used.
For instance, depending on the situation, the word “luck” can be used either posi-
tively or negatively.
– Grammatical ambiguities: The sentiment of a phrase may depend on the gram-
mar used. For example, the phrase “it isn’t bad” may be interpreted as having a
positive sentiment, whereas the phrase “it isn’t good” may be interpreted as hav-
ing a negative sentiment.
Linear algorithms are used in conventional approaches to sentiment analysis. This often
entails using a machine learning technique to classify text as positive or negative, such
as a support vector machine or linear regression [7]. These methods frequently depend
on the usage of NLP features such part-of-speech tags, word embeddings, and sentiment
lexicons and call for a large training data collection of labeled instances. With simple
language and more conventional use cases, these methods typically work well.
348 M. V. Jagannatha Reddy et al.
Extraction of usable information from vast volumes of data is known as data mining.
To find patterns, trends, and linkages, massive datasets must be sorted through. Tech-
niques like machine learning, AI, and statistics are used in data mining to evaluate
data and make inferences from it [9]. Data mining can be used to make predictions
about future trends by understanding current patterns. Data mining is used in many
areas such as healthcare, finance, marketing, and education. It can be used to gather
information about customer preferences, and to analyze customer buying patterns.
Data mining is also used to identify fraud and other anomalies in datasets.
Gathering, cleaning, organizing, and transforming data into a format that can be used
by a machine learning algorithm for modeling and analysis are all parts of data collec-
tion and preprocessing. Data collection typically involves obtaining information from
a variety of sources, including surveys, databases, websites, and social media [10].
Data preprocessing involves cleaning the data to correct errors, fixing inconsis-
tencies, reducing noise and outliers, and transforming it into a form suitable for use
by a machine learning algorithm. This process includes normalizing the data, trans-
forming it into a suitable form, and eliminating redundant information. Additionally,
data preprocessing often involves reducing the dimensions of the dataset by combin-
ing or eliminating features that are not relevant to the problem.
17 Exploring artificial intelligence techniques for enhanced sentiment analysis 349
The technique of separating useful information from unprocessed data is called fea-
ture extraction. By lowering the amount of characteristics that describe the data, it is
used to reduce the dimensionality of datasets. The process of choosing which features
to include in the modeling process is known as feature selection [11]. The method may
assist in increasing the model’s correctness or decreasing the computational expense
associated with model evaluation. A subset of traits that have the greatest influence
on a desired outcome are sought after through feature selection. The type of feature
selection will depend largely on the type of data being analyzed. For example, a data-
set consisting of text documents would require a different type of feature extraction/
selection than a dataset consisting of images.
The process of evaluating a text to ascertain the underlying sentiment of the tone is
known as sentiment analysis. In other words, it involves computationally recognizing
and classifying viewpoints that people express in written form. In order to learn the
sentiment related to a certain text, machine learning algorithms for sentiment analy-
sis often rely on supervised learning approaches [12]. These algorithms often use la-
beled training data, which teaches the algorithm to associate the sentiment of certain
words or phrases with the overall sentiment of the text. The trained model can then
be used by the computer to ascertain the sentiment of new, unexplored material.
Naive Bayes, Support Vector Machines, and Decision Trees are common algorithms
for sentiment analysis.
The accuracy of a sentiment analysis model is measured using evaluation metrics for
sentiment analysis models. The performance of the model on a particular set of data
is often quantified by these metrics, which may be used to compare other models. The
area under the curve (AUC), F1, precision, recall, and accuracy are typical measures
employed for this purpose. Additional metrics include the less popular Matthews cor-
relation coefficient (MCC), mean absolute error, and mean square error. All of these
metrics measure the model’s ability to correctly detect sentiment in texts, and each
metric offers a different measure [13].
350 M. V. Jagannatha Reddy et al.
Understanding and processing human language is the focus of the branch of AI known
as NLP. It makes it possible for machines to meaningfully analyze, comprehend, and pro-
duce natural language. The goal of sentiment analysis, which employs NLP, is to locate
and classify the opinions stated in a text, particularly those relating to a good or service.
It is used to identify client emotions, such as happiness, rage, or contempt toward a spe-
cific subject or product [16]. In order to analyze a text’s sentiment, compare it to estab-
lished categories, and assign a sentiment score, NLP uses machine learning models like
naive Bayes, support vector machine, and RNNs. This score can then be used to deter-
mine customer’s opinion about a particular product or service.
17 Exploring artificial intelligence techniques for enhanced sentiment analysis 351
Deep learning models create sentiment analysis for text documents using neural net-
works. These advanced networks are able to learn underlying features and patterns
in text documents that classify them as positive or negative. This allows deep learning
models to detect even subtle nuances in sentiment that human readers may miss [17].
Deep learning models recognize complex linguistic patterns, such as sarcasm, and can
even recognize when an opinion is being expressed by an individual. These models
can be trained on large datasets, allowing them to glean more information from a sin-
gle text sample. As a result, deep learning-based sentiment analysis is known to be
more accurate than traditional machine learning methods.
Transfer learning is the process of using sentiment analysis models that have already
been trained. To solve a new problem, it entails using a pretrained model and fine-
tuning its weights on a fresh dataset. This method can greatly increase a model’s accu-
racy and performance while reducing the amount of labeled data required to train a
new model. It has been used to increase the precision of sentiment analysis models in
machine learning and deep learning [18]. In sentiment analysis, transfer learning can
be used to train a model on a dataset of labeled reviews from a specific domain and
then use that model to analyze reviews in a different domain. This saves time and
resources since you are not training a model from scratch for each new domain.
A potent method for determining the sentiment of text is to use word embeddings for
sentiment analysis. Word embeddings are numerical representations of words, capa-
ble of capturing semantic relationships between the words. Sentiment analysis is the
understanding of a text’s emotional tone, expressed through words or phrases, to de-
termine an overall sentiment. Utilizing word embeddings in sentiment analysis ena-
bles the capturing of associations between words; for instance, if a word is linked
with a positive sentiment, it may alternatively be connected with a negative sentiment
if it appears next another word that is associated with a negative attitude [22]. This
contextual relationship between words is what enables word embeddings to capture
sentiment. By leveraging word embeddings, sentiment analysis can be achieved faster
and more accurately than using traditional methods, such as bag-of-words.
Negation and sarcasm in sentiment analysis can be extremely tricky to detect as they
are highly subjective and sometimes depend on tone, context, and even cultural back-
ground. Negation occurs when a negative word is used, or when the opinion ex-
pressed using positive words is intended to have a negative connotation. For example,
17 Exploring artificial intelligence techniques for enhanced sentiment analysis 353
the phrase “not bad” could indicate something that is either acceptable or mediocre,
depending on how it is said. Sarcasm is a form of irony which often requires an un-
derstanding of cultural context and past events to be detected accurately [23]. It often
uses positive language or phrases to portray a negative sentiment. Sentiment analysis
requires more than just a vocabulary of positive and negative terms to handle nega-
tion and sarcasm. Irony and sarcasm can be recognized in language by using NLP
techniques. Natural language understanding (NLU) can be used to analyze the context
of a statement to infer the intended sentiment. Machine Learning algorithms can be
employed to detect sentiment in a large amount of data. This can involve training the
models to recognize nuances like irony, sarcasm, or negation. Ultimately, sentiment
analysis tools should rely on both human expertise and AI approaches to accurately
capture the sentiment of a statement.
Imbalanced data is an issue in sentiment analysis as it affects the accuracy of the clas-
sification algorithms. Class imbalances in sentiment datasets arise from a minority of
sentiment classes being heavily present in the dataset, while other sentiment classes
are minority classes, and often difficult to classify accurately. Addressing class imbal-
ance in sentiment datasets is important to ensure accurate predictions for all senti-
ment classes [25]. Undersampling or oversampling data points from the minority
354 M. V. Jagannatha Reddy et al.
classes, or combining the two strategies, is the most typical method of addressing
class imbalance in emotion datasets. When sampling is too small, some data points
from the majority class are discarded. Undersampling is used to increase the amount
of data points from the minority class while decreasing the number of data points
from the majority class. This helps to create a more balanced dataset that can be used
to train the sentiment classifier [26]. Oversampling involves the duplication of minor-
ity data points. This technique can be used to create more training data for the minor-
ity class. Oversampling is used to provide a more balanced dataset for training the
sentiment classifier by increasing the quantity of data points from the minority class.
In addition to undersampling and oversampling, another technique that can be used
to address class imbalance in sentiment datasets is to introduce weights to the data
points in different classes. Weights can be used to increase the importance of minority
classes during training and reduce the importance of majority classes. To sum up, the
techniques that can be used to address class imbalance in sentiment datasets are
undersampling, oversampling, and introducing weights to the data points in different
classes. These approaches can help to ensure accurate predictions for all sentiment
classes.
Sentiment analysis in social media is the process of identifying and extracting subjec-
tive information from user-generated material in social media using NLP and computa-
tional linguistic approaches. It is employed to track public opinion about a certain
company, good, or service. By categorizing comments made on social media as favor-
able, negative, or neutral, it aids in determining how the general population feels about
a particular subject [29]. For sentiment categorization, topic extraction, opinion mining,
sentiment trend analysis, and brand monitoring, sentiment analysis can be employed.
Customer reviews about a product, service, or company are analyzed using a process
called sentiment analysis. By recognizing any positive, negative, or neutral sentiment
that may be present, sentiment analysis seeks to quantify the total sentiment of cus-
tomer reviews. This can be used to identify trends in customer sentiment and gain an
understanding of how customers feel about the product, service, or business in ques-
tion [30]. The output of such an analysis can be used to inform marketing or product
strategy decisions.
Challenges:
1. Natural language understanding: Sentiment analysis is based on NLU technology,
which is still far from a perfect science. Complex linguistic constructs, slang, acro-
nyms, etc. can all throw off otherwise perfectly accurate sentiment analysis.
2. Context and subjectivity: Sentiment analysis is subjective and depends on the con-
text of a given message. Sentiment analysis is susceptible to context misunder-
standing because the context in which a sentence is used might modify its
meaning.
3. Sentiment polarity: The polarity of a statement is difficult to detect, given the sub-
jectivity involved. Many sentiment analysis algorithms use a binary classification
such as positive vs. negative but there is a need for a more nuanced approach
that can better identify sentiment polarity across a wider range of emotions.
Future directions:
1. Deep learning: A subclass of AI, deep learning is particularly helpful for challeng-
ing tasks like sentiment analysis. Deep learning neural networks are capable of
discovering patterns in data that are not accessible to conventional machine
learning algorithms. This can result in sentiment analysis that is more precise.
2. Transfer learning: This method can be applied to increase the precision of senti-
ment analysis. The accuracy of sentiment analysis can be increased by utilizing
pretrained models developed on substantial datasets.
3. Social media analysis: As the use of social media keeps expanding, sentiment
analysis of social media data is an area of growing interest. Decisions can be
made using social media sentiment analysis, which can offer insightful informa-
tion about public opinion on a variety of subjects.
– The current AI techniques for sentiment analysis have difficulty handling sar-
casm, irony, and nuance as sentiment is not always clearly stated.
– AI sentiment analysis algorithms are not yet able to truly understand the true
sentiment of a statement or phrase, as a result of the lack of ability to infer mean-
ing and context.
– The restricted capacity of AI sentiment analysis to obtain and examine vast vol-
umes of organized and unstructured data is another drawback.
– Many current AI models have difficulty recognizing and understanding the intent
of a statement.
17 Exploring artificial intelligence techniques for enhanced sentiment analysis 357
Ethical considerations in sentiment analysis refer to the moral dilemmas that may
arise when analyzing the sentiment of text or objects [32]. It includes understanding
how people’s ideologies, cultures, and even personal beliefs may influence the analy-
sis. As sentiment analysis becomes increasingly popular, it is important to adhere and
consider ethical practices. This includes respecting people’s privacy, protecting confi-
dential data, and avoiding bias when analyzing data. Additionally, it is important to
consider how sentiment analysis can affect certain groups, such as those belonging to
minority communities or even those belonging to technological literacies [33]. Finally,
the ethical considerations include the understanding of the implications of the, mak-
ing sure they are considered when drawing conclusions from the analysis.
In the future, AI will continue to be used in sentiment analysis, but the way it is used
will be more refined and specialized. AI-powered sentiment analysis will become
more accurate and adept at interpreting the complexities of language and subtle nu-
ances in sentiment [34, 35]. In order to better comprehend emotion, AI will also be
utilized to create better datasets and feature sets. NLP, which uses AI to comprehend
the meaning of words and phrases, is another increasingly significant method for sen-
timent analysis. AI techniques such as deep learning and neural networks are being
used to identify sentiment in texts more accurately and faster than ever before. In
addition, AI-powered sentiment analysis applications are becoming increasingly more
context aware. By utilizing NLP, AI can interpret the nuanced emotions associated
with a particular sentence or phrase and interpret the sentiment accurately [36]. The
AI-driven sentiment analysis will be used to better understand customer behavior. AI
can analyze customer reviews and conversations from various platforms and provide
businesses with valuable insights on how customers interact with and feel about dif-
ferent products or services. This insight can then be used to create better customer
experiences and improve customer service.
358 M. V. Jagannatha Reddy et al.
References
[1] Xu, G., Meng, Y., Qiu, X., Yu, Z., & Wu, X. (2019). Sentiment analysis of comment texts based on
BiLSTM. IEEE Access, 7, 51522–51532.
[2] Mehta, P., & Pandya, S. (2020). A review on sentiment analysis methodologies, practices and
applications. International Journal of Scientific and Technology Research, 9(2), 601–609.
[3] Zad, S., Heidari, M., Jones, J. H., & Uzuner, O. (2021, May). A survey on concept-level sentiment
analysis techniques of textual data. In 2021 IEEE World AI IoT Congress (AIIoT) (pp. 0285–0291).
IEEE.
[4] Pereira, D. A. (2021). A survey of sentiment analysis in the Portuguese language. Artificial
Intelligence Review, 54(2), 1087–1115.
[5] Elfaik, H., & Nfaoui, E. H. (2020). Deep bidirectional LSTM network learning-based sentiment analysis
for Arabic text. Journal of Intelligent Systems, 30(1), 395–412.
[6] Chen, L. C., Lee, C. M., & Chen, M. Y. (2020). Exploration of social media for sentiment analysis using
deep learning. Soft Computing, 24, 8187–8197.
[7] Saura, J. R., Palos-Sanchez, P., & Grilo, A. (2019). Detecting indicators for startup business success:
Sentiment analysis using text data mining. Sustainability, 11(3), 917.
[8] Mittal, A., & Patidar, S. (2019, July). Sentiment analysis on twitter data: A survey. In Proceedings of
the 7th International Conference on Computer and Communications Management (pp. 91–95).
[9] Kurani, A., Doshi, P., Vakharia, A., & Shah, M. (2023). A comprehensive comparative study of artificial
neural network (ANN) and support vector machines (SVM) on stock forecasting. Annals of Data
Science, 10(1), 183–208.
[10] Basiri, M. E., Nemati, S., Abdar, M., Cambria, E., & Acharya, U. R. (2021). ABCDM: An attention-based
bidirectional CNN-RNN deep model for sentiment analysis. Future Generation Computer Systems,
115, 279–294.
[11] Wankhade, M., Rao, A. C. S., & Kulkarni, C. (2022). A survey on sentiment analysis methods,
applications, and challenges. Artificial Intelligence Review, 55(7), 5731–5780.
[12] Ali, N. M., Abd El Hamid, M. M., & Youssif, A. (2019). Sentiment analysis for movies reviews dataset
using deep learning models. International Journal of Data Mining & Knowledge Management
Process (IJDKP), 1–9.
[13] Yadav, S. P., & Yadav, S. (2018). Fusion of medical images in wavelet domain: a discrete
mathematical model. Ingeniería Solidaria, 14(25), 1–11.
[14] Gobinathan, B., Mukunthan, M. A., Surendran, S., Somasundaram, K., Moeed, S. A., Niranjan, P., . . .
& Sundramurthy, V. P. (2021). A novel method to solve real time security issues in software industry
using advanced cryptographic techniques. Scientific Programming, 2021, 1–9.
[15] Yadav, S. P., & Yadav, S. (2019). Mathematical implementation of fusion of medical images in
continuous wavelet domain. Journal of Advanced Research in Dynamical and Control System, 10(10),
45–54.
[16] Sivaram, M., Yuvaraj, D., Mohammed, A. S., Manikandan, V., Porkodi, V., & Yuvaraj, N. (2019).
Improved enhanced DBTMA with contention-aware admission control to improve the network
performance in MANETS. CMC-Computers Materials and Continua, 60(2), 435–454.
[17] Rani, P., Verma, S., Yadav, S. P., Rai, B. K., Naruka, M. S., & Kumar, D. (2022). Simulation of the
lightweight blockchain technique based on privacy and security for healthcare data for the cloud
system. International Journal of E-Health and Medical Communications (IJEHMC), 13(4), 1–15.
[18] Manikandan, V., & Yuvaraj, N. (2019, May). Advanced expert system using particle swarm
optimization based adaptive network based fuzzy inference system to diagnose the physical
constitution of human body. In Emerging Technologies in Computer Engineering: Microservices in
Big Data Analytics: Second International Conference, ICETCE 2019, Jaipur, India, February 1–2, 2019,
Revised Selected Papers (Vol. 985, pp. 349). Springer.
17 Exploring artificial intelligence techniques for enhanced sentiment analysis 359
[19] Pujahari, R. M., Yadav, S. P., & Khan, R. (2022). Intelligent farming system through weather forecast
support and crop production. In : Mohammad Ayoub Khan, Rijwan Khan and Mohammad Aslam
Ansari (Eds.). Application of Machine Learning in Agriculture (pp. 113–130). United States: Academic
Press.
[20] Yuvaraj, N., Raja, R. A., Kousik, N. V., Johri, P., & Diván, M. J. (2020). Analysis on the prediction of
central line-associated bloodstream infections (CLABSI) using deep neural network classification. In:
Jitendra Kumar Verma, Sudip Paul and Prashant Johri (Eds.). Computational Intelligence and Its
Applications in Healthcare (pp. 229–244). United States: Academic Press.
[21] Irfan, D., Tang, X., Narayan, V., Mall, P. K., Srivastava, S., & Saravanan, V. (2022). Prediction of Quality
Food Sale in Mart Using the AI-Based TOR Method. Journal of Food Quality, 2022, 1–13.
[22] Ramesh, G., Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). Prediction of energy production level
in large PV plants through AUTO-Encoder Based Neural-Network (AUTO-NN) with restricted
Boltzmann feature extraction. Future Internet, 15(2), 46.
[23] Rajalakshmi, M., Saravanan, V., Arunprasad, V., Romero, C. T., Khalaf, O. I., & Karthik, C. (2022).
Machine learning for modeling and control of industrial clarifier process. Intelligent Automation
and Soft Computing, 32(1), 339–359.
[24] Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model
for efficient resource allocation in 5G broadband communication networks. ICTACT Journal on Soft
Computing, 13(3), 2986–2994.
[25] Saravanan, V., & Neeraja, A. (2013, January). Security issues in computer networks and
stegnography. In 2013 7th International Conference on Intelligent Systems and Control (ISCO)
(pp. 363–366). IEEE.
[26] Sangeetha, R., Logeshwaran, J., Rocher, J., & Lloret, J. (2023). An improved agro deep learning model
for detection of Panama wilts disease in banana leaves. AgriEngineering, 5(2), 660–679.
[27] Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J., . . . &
Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey
wolf optimization in wireless sensor networks. Security and Communication Networks, 2022, 1–13.
[28] Logeshwaran, J., Kiruthiga, T., Kannadasan, R., Vijayaraja, L., Alqahtani, A., Alqahtani, N., & Alsulami,
A. A. (2023). Smart load-based resource optimization model to enhance the performance of device-
to-device communication in 5G-WPAN. Electronics, 12(8), 1821.
[29] Harish, B. S., Kumar, K., & Darshan, H. K. (2019). Sentiment analysis on IMDb movie reviews using
hybrid feature extraction method. International Journal of Interactive Multimedia and Artificial
Intelligence, 45(2), 1–10.
[30] Jain, P. K., Pamula, R., & Srivastava, G. (2021). A systematic literature review on machine learning
applications for consumer sentiment analysis using online reviews. Computer Science Review, 41,
100413.
[31] Chen, X., Zou, D., Xie, H., Cheng, G., & Liu, C. (2022). Two decades of artificial intelligence in
education. Educational Technology and Society, 25(1), 28–47.
[32] Abualigah, L., Alfar, H. E., Shehab, M., & Hussein, A. M. A. (2020). Sentiment analyss in healthcare: A
brief review. Recent Advances in NLP: The Case of Arabic Language, 67, 129–141.
[33] Kumar, A., & Garg, G. (2020). Systematic literature review on context-based sentiment analysis in
social multimedia. Multimedia Tools and Applications, 79, 15349–15380.
[34] Alsaeedi, A., & Khan, M. Z. (2019). A study on sentiment analysis techniques of Twitter data.
International Journal of Advanced Computer Science and Applications, 10(2), 1–18.
[35] Xu, G., Yu, Z., Yao, H., Li, F., Meng, Y., & Wu, X. (2019). Chinese text sentiment analysis based on
extended sentiment dictionary. IEEE Access, 7, 43749–43762.
[36] Nemes, L., & Kiss, A. (2021). Social media sentiment analysis based on COVID-19. Journal of
Information and Telecommunication, 5(1), 1–15.
Senthil Kumar C, A. R. Arunarani, Piyush Charan,
Sanjeev Kumar Angadi
18 Exploring the potential of artificial
intelligence for automated sentiment
analysis and opinion mining
Abstract: The development of automated sentiment analysis and opinion mining ap-
proaches for enhancing the precision and scalability of text analysis has been made
possible by the rising integration of artificial intelligence (AI) technology in digital sys-
tems. The method of automatically evaluating and extracting sentiment-related data
from a text is known as automatic sentiment analysis. Automatically extracting and
summarizing opinions from a text is referred to as opinion mining. AI has enormous
promise for automated sentiment analysis and opinion mining, which can increase
consumer engagement and experience.
18.1 Introduction
Artificial intelligence (AI)-based automated sentiment analysis and opinion mining
techniques provide a reliable means of deriving customers’ emotion or opinions from
text sources such as social posts, reviews, emails, and surveys. Such technique can
help to uncover customer sentiment on a product, service, or company, allowing busi-
nesses to better understand their customers in terms of overall satisfaction and pro-
vide better customer service [1]. AI-driven techniques further allow for automated
processes for deriving the sentiment of large numbers of text sources. This is espe-
cially beneficial in industries where customer feedback is essential, such as in cus-
tomer service, healthcare, retail, and hospitality.
Automated sentiment analysis and AI have completely changed how people interact
and communicate online. Machines can now comprehend and further extract mean-
ing from text thanks to AI technologies like natural language processing (NLP) [2]. In
turn, this has made it possible for automated sentiment analysis and opinion mining
techniques to better understand people’s sentiments and ideas as they are conveyed
in textual data.
https://doi.org/10.1515/9783111323749-018
362 Senthil Kumar C et al.
The objectives of AI for automated sentiment analysis and opinion mining are:
1. To understand human emotions in social media and other textual data.
2. To identify the opinions of the author about any product or service by analyzing
the text in the review.
3. To classify sentiment into one of the four categories (positive, negative, neutral,
or mixed) based on the content of the review.
4. Identifying trends in opinion for further market research, analysis, and decision-
making.
5. Identifying customer service issues and providing timely solutions.
6. Encouraging customer engagement and loyalty.
7. To generate insights on customer sentiment to make improvements to product
and services.
8. To predict customer behavior and analyze the impact of marketing activities.
The usage of automated systems has a significant impact on the capabilities and restric-
tions of AI for automated sentiment analysis and opinion mining. Text analysis techni-
ques are typically used to discover the emotional content in texts in order for automated
sentiment analysis and opinion mining to be able to determine the overall sentiment
communicated in the text. Additionally, it can identify the precise opinions contained in
the text, as well as the mood of those ideas (whether good or negative) [5].
18 Exploring the potential of artificial 363
However, automated sentiment analysis and opinion mining have some limita-
tions. They are often reliant upon NLP techniques, which means the accuracy of the
outputs are greatly influenced by the complexity of the text. Automated systems are
also limited in their ability to understand the sentiment and opinion being conveyed
in a text. For example, the systems might look for certain key phrases or words to
detect sentiment, but may fail to identify the more nuanced elements of sentiment,
such as sarcasm or double-entendres. The quality of the data being used in the analy-
sis will also affect automated sentiment analysis and opinion mining. Erroneous or
biased data in a dataset may produce erroneous results. It is crucial to be aware of
the potential scope and restrictions when establishing an automated system because
automated sentiment analysis and opinion mining have enormous potential to be em-
ployed in a variety of applications [6].
Identifying the emotional undertone of a string of words through the method of senti-
ment analysis allows one to classify a passage of text as positive, negative, or neutral.
Finding out about the attitudes, ideas, and feelings conveyed in the text can be accom-
plished through studying sentiment in the text. This can be used to identify trends in
public opinion or understand consumer sentiment towards a particular product, ser-
vice, or company [8]. It can also be used to detect sarcasm and other subtleties in lan-
guage that are not able to be detected through traditional textual analysis methods.
The method of sentiment analysis involves employing text analysis technologies to au-
tomatically identify the conveyed sentiment in a text. The process, commonly referred
364 Senthil Kumar C et al.
Challenges:
– Interpretability: As opinion mining techniques are mainly based on complex algo-
rithms, it can be difficult to understand their internal behavior. This makes it
hard to build trust in the opinion mining systems and interpret the results they
generate accurately.
– Contextual Issues: Opinion mining systems can be affected by context-specific
terms and phrases, which can make it difficult to generalize findings and apply
them in similar situations.
– Ambiguity: The meanings of words can differ based on the context and author,
making it difficult to understand the true sentiment behind a particular phrase or
sentence.
– Variety of opinions: Opinion mining systems are often trained on a limited data-
set, which does not always reflect the full range of opinions that exist in the real
world.
Limitations:
– Limited accuracy: The accuracy of opinion mining systems can still be affected by
the quality of the dataset they are trained on and the complexity of the algorithms.
– Expensive: Developing and deploying a robust opinion mining system can be
costly, as it requires a significant amount of resources in terms of data acquisi-
tion, development, and deployment.
– Time-consuming: Training an opinion mining system can be time-consuming and
requires expertise in machine learning. This can lead to delayed results in a live
system.
18 Exploring the potential of artificial 365
One of the most debated and cutting-edge technologies used in enterprises today is AI.
Robotic process automation (RPA), deep learning, machine learning (ML), and NLP
are just a few of the methods used in AI. AI is designed to allow machines and com-
puter programs to learn from experience, identify trends, and make decisions with
human-like intelligence. AI techniques are applied to enterprise applications such as
security systems, rules engines, chatbots, and enterprise search to automate pro-
cesses, provide higher levels of accuracy, and reduce costs across many industries
[13]. AI is also becoming more widely used in the healthcare sector to identify and
manage patient health conditions, reduce medical errors, and assist with precise drug
doses and medication management. Furthermore, AI is used to power autonomous
vehicles as well as to power robotic arms in modern factories. AI is constantly advanc-
ing and being used in a greater number of applications, and more innovative uses are
continually being developed.
In ML, a subset of AI, computers may “learn” from data without having to be explicitly
taught. In order to produce predictions or choices, ML approaches rely on algorithms
that can spot patterns in data. ML approaches can be divided into three categories:
supervised learning, unsupervised learning, and reinforcement learning [14].
Algorithms are trained on data and given the desired outcomes as part of super-
vised learning. The algorithm will look for a pattern that corresponds to the antici-
pated outcomes as new data is submitted to it. You can utilize supervised learning to
find solutions to regression, classification, and ranking issues.
Contrarily, unsupervised learning doesn’t rely on the data having labels that have
already been established. Instead, the system is given an unlabeled dataset and is
366 Senthil Kumar C et al.
forced to identify patterns on its own. Clustering and anomaly detection are two com-
mon unsupervised learning applications.
A further ML strategy is the use of reinforcement learning algorithms. Algorithms
for reinforcement learning are taught to iterate over challenges to find the optimal
answer. Problems requiring the agent to learn from its environment or have complex
incentive systems can be resolved with this type of learning.
The machine is taught about the connections between input and output data via super-
vised learning, a sort of ML method. By giving the algorithm labeled examples to refer
to when making judgments and predictions, the system gets taught. A prediction model
is created using the input data using the supervised learning algorithm. After that, this
model is used to unobserved data to produce predictions and categorizations [15].
A ML method called semi-supervised learning trains models using both labeled and
unlabeled input. The main goal of semi-supervised learning is to supplement labeled
data with unlabeled data in order to learn from a bigger set of data. This will help the
model generalize more effectively and enhance its predictive ability. Semi-supervised
learning algorithms are typically used for classification and clustering tasks [17].
Some example scenarios of using semi-supervised learning include image recognition,
spam detection, and NLP.
18 Exploring the potential of artificial 367
A sort of artificial neural network called a convolutional neural network (CNN) is em-
ployed in computer vision, NLP, and a number of other applications. A CNN functions
by taking a set of input images, running a series of convolutional filters across the
images at various sizes, and then extracting a set of features from the images. The
tasks that the features from the CNN may subsequently be applied to include senti-
ment analysis, object recognition, and image segmentation [20]. Typically, CNNs oper-
ate by taking a set of input images, running a number of convolutional filters across
the images at various scales, pooling the features, and then flattening the input im-
ages into a high-dimensional vector.
368 Senthil Kumar C et al.
A unique kind of RNN called long short-term memory (LSTM) is utilized in deep learn-
ing. It is used to remember long-term data dependencies and is made for sequence
prediction challenges. Using unique “memory cells” that can retain data for lengthy
periods of time, it can solve the vanishing gradient problem experienced by conven-
tional RNNs [22]. It can accurately detect patterns in data sequences as well as learn
complex connections. LSTM can be applied to a variety of tasks, including time series
analysis, text generation, language modeling, and speech recognition.
Rule-based approaches in sentiment analysis use predefined rules like regular expres-
sions, lexicons, and syntactic parsing to classify the sentiment of text. The goal of
these methods is to classify the sentiment of text by using predefined rules [24].
These methods rely heavily on prior knowledge of language and can be used out
of the box with a set of predefined rules. However, for a particular domain or lan-
guage, custom setup rules are also necessary for the successful application of the rule-
based approach [25].
The main advantages of rule-based approaches in sentiment analysis are that
they are easy to understand, easy to implement, and cost-effective. Compared to ML-
based approaches, they require less maintenance and data processing as pre-defined
rules are used. On the downside, rule-based approaches are limited in their ability to
detect complex sentiment and associations between words [26].
Lexicon-based approaches for sentimental analysis are a type of NLP technique that
use dictionary-based lexicons to analyze the sentiment of a given text. This method
entails labeling words in a text with sentimental categories like positive, negative, or
neutral. After that, the document’s overall sentiment is ascertained using the senti-
ment labels. In lexicon-based methods, sentiment labels are applied to specific words
in accordance with a specified sentiment lexicon. This lexicon is often a set of words
and phrases with the labels that correspond to the respective sentiments [27]. With
this method, sentiment analysis may be completed for vast amounts of text rapidly
and accurately. Due to the fact that the sentiment labels have already been estab-
lished, little training data is also needed. Last but not least, this strategy may typically
be interpreted in terms of the underlying sentiment labels employed in the sentiment
lexicon.
ment analysis aids in determining the sentiment of a text [28]. For instance, it is possi-
ble to evaluate a movie review to ascertain the general mood – whether positive or
negative – as well as the reviewer’s opinions on particular elements, such as the cast,
direction, and script. Aside from helping with customer support, ABSA can also be
used to learn more about how customers feel about a given good or service.
References
[1] Yadav, S. P., Zaidi, S., Mishra, A., & Yadav, V. (2022). Survey on machine learning in speech emotion
recognition and vision systems using a recurrent neural network (RNN). Archives of Computational
Methods in Engineering, 29(3), 1753–1770.
[2] Khan, S., Saravanan, V. N. G. C., Lakshmi, T. J., Deb, N., & Othman, N. A. (2022). Privacy protection of
healthcare data over social networks using machine learning algorithms. Computational
Intelligence and Neuroscience, 2022, 1–12.
[3] Yadav, S. P. (2021). Emotion recognition model based on facial expressions. Multimedia Tools and
Applications, 80(17), 26357–26379.
[4] Hewahi, N. M., & Baraka, A. M. (2012). Emotion recognition model based on facial expressions,
ethnicity and gender using backpropagation neural network. International Journal of Technology
Diffusion (IJTD), 3(1), 33–43.
[5] Gopi, B., Logeshwaran, J., Gowri, J., & Kiruthiga, T. (2022). The moment probability and impacts
monitoring for electron cloud behavior of electronic computers by using quantum deep learning
model. NeuroQuantology, 20(8), 6088–6100.
[6] Munagala, N. V. L., Saravanan, V., Almukhtar, F. H., Jhamat, N., Kafi, N., & Khan, S. (2022). Supervised
approach to identify autism spectrum neurological disorder via label distribution learning.
Computational Intelligence and Neuroscience, 2022, 1–9.
[7] Yadav, S. P., Mahato, D. P., & Linh, N. T. D. (Eds.). (2020). Distributed artificial intelligence: A modern
approach. Boca Raton: CRC Press.
[8] Raja, S., Logeshwaran, J., Venkatasubramanian, S., Jayalakshmi, M., Rajeswari, N., Olaiya, N. G., &
Mammo, W. D. (2022). OCHSA: Designing Energy-Efficient Lifetime-Aware Leisure Degree Adaptive
Routing Protocol with Optimal Cluster Head Selection for 5G Communication Network Disaster
Management. Scientific Programming, 2022.
[9] Saravanan, V., & Mohan Raj, V. (2016). Maximizing QoS by cooperative vertical and horizontal
handoff for tightly coupled WIMAX/WLAN overlay networks. Cluster Computing, 19(3), 1619–1633.
[10] Yuvaraj, N., Karthikeyan, T., & Praghash, K. (2021). An improved task allocation scheme in serverless
computing using gray wolf optimization (GWO) based reinforcement learning (RIL) approach.
Wireless Personal Communications, 117, 2403–2421.
[11] Sumathi, A., & Saravanan, V. (2015). Bandwidth based vertical handoff for tightly coupled wimax/
wlan overlay networks. Journal of Scientific and Industrial Research, 74(10), 560–566.
[12] Logeshwaran, J., Malik, J. A., Adhikari, N., Joshi, S. S., & Bishnoi, P. (2022). IoT-TPMS: An innovation
development of triangular patient monitoring system using medical internet of things. International
Journal of Health Sciences, 6(S5), 9070–9084.
[13] Sivaram, M., Porkodi, V., Mohammed, A. S., Manikandan, V., & Yuvaraj, N. (2019). Retransmission
DBTMA protocol with fast retransmission strategy to improve the performance of manets. IEEE
Access, 7, 85098–85109.
[14] Yadav, S. P. (2020). Vision-based detection, tracking, and classification of vehicles. IEIE Transactions
on Smart Processing and Computing, 9(6), 427–434.
[15] Kousik, N., Natarajan, Y., Raja, R. A., Kallam, S., Patan, R., & Gandomi, A. H. (2021). Improved salient
object detection using hybrid Convolution Recurrent Neural Network. Expert Systems with
Applications, 166, 114064.
[16] Yuvaraj, N., & Vivekanandan, P. (2013, February). An efficient SVM based tumor classification with
symmetry non-negative matrix factorization using gene expression data. In 2013 International
Conference on Information Communication and Embedded Systems (ICICES) (pp. 761–768). IEEE.
[17] Ramesh, G., Logeshwaran, J., & Rajkumar, K. (2022). The smart construction for image preprocessing
of mobile robotic systems using neuro fuzzy logical system approach. NeuroQuantology, 20(10),
6354–6367.
372 Senthil Kumar C et al.
[18] Kaur, A., & Gupta, V. (2013). A survey on sentiment analysis and opinion mining techniques. Journal
of Emerging Technologies in Web Intelligence, 5(4), 367–371.
[19] Osimo, D., & Mureddu, F. (2012). Research challenge on opinion mining and sentiment analysis.
Universite de Paris-Sud, Laboratoire LIMSI-CNRS, Bâtiment, 508, 1–13.
[20] Vinodhini, G., & Chandrasekaran, R. M. (2012). Sentiment analysis and opinion mining: A survey.
International Journal, 2(6), 282–292.
[21] Cambria, E., Schuller, B., Xia, Y., & Havasi, C. (2013). New avenues in opinion mining and sentiment
analysis. IEEE Intelligent Systems, 28(2), 15–21.
[22] Chen, H., & Zimbra, D. (2010). AI and opinion mining. IEEE Intelligent Systems, 25(3), 74–80.
[23] Ravi, K., & Ravi, V. (2015). A survey on opinion mining and sentiment analysis: Tasks, approaches
and applications. Knowledge-based Systems, 89, 14–46.
[24] Alexandridis, G., Varlamis, I., Korovesis, K., Caridakis, G., & Tsantilas, P. (2021). A survey on sentiment
analysis and opinion mining in greek social media. Information, 12(8), 331.
[25] Othman, M., Hassan, H., Moawad, R., & El-Korany, A. (2014). Opinion mining and sentimental
analysis approaches: A survey. Life Science Journal, 11(4), 321–326.
[26] Aquino, P. A., López, V. F., Moreno, M. N., Muñoz, M. D., & Rodríguez, S. (2020). Opinion mining
system for Twitter sentiment analysis. In Hybrid Artificial Intelligent Systems: 15th International
Conference, HAIS 2020, Gijón, Spain, November 11–13, 2020, Proceedings (Vol. 15, pp. 465–476).
Springer International Publishing.
[27] Liu, B., & Zhang, L. (2012). A survey of opinion mining and sentiment analysis. In: Charu C. Aggarwal,
ChengXiang Zhai (Eds.). Mining Text Data (pp. 415–463). Boston, MA: Springer.
[28] Keikhosrokiani, P., & Pourya Asl, M. (Eds.). (2022). Handbook of research on opinion mining and text
analytics on literary works and social media. United States: IGI Global.
[29] Lula, P., & Wójcik, K. (2011). Sentiment analysis of consumer opinions written in Polish. Economics
and Management, 16(1), 1286–1291.
[30] Navale, G. S., Dudhwala, N., Jadhav, K., Gabda, P., & Vihangam, B. K. (2016). Prediction of stock
market using data mining and artificial intelligence. International Journal of Computer Applications,
134(12), 9–11.
[31] Pang, B., & Lee, L. (2008). Opinion mining and sentiment analysis. Foundations and Trends® in
Information Retrieval, 2(1–2), 1–135.
[32] Haddi, E., Liu, X., & Shi, Y. (2013). The role of text pre-processing in sentiment analysis. Procedia
Computer Science, 17, 26–32.
[33] Kirilenko, A. P., Stepchenkova, S. O., Kim, H., & Li, X. (2018). Automated sentiment analysis in
tourism: Comparison of approaches. Journal of Travel Research, 57(8), 1012–1025.
[34] Zad, S., Heidari, M., Jones, J. H., & Uzuner, O. (2021, May). A survey on concept-level sentiment
analysis techniques of textual data. In 2021 IEEE World AI IoT Congress (AIIoT) (pp. 0285–0291).
New York City: IEEE.
[35] Banker, S., & Patel, R. (2016). A brief review of sentiment analysis methods. International Journal of
Information Sciences and Techniques (IJIST), 6(1/2), 89–95.
[36] Grljević, O., Bošnjak, Z., & Kovačević, A. (2022). Opinion mining in higher education: A corpus-based
approach. Enterprise Information Systems, 16(5), 1773542.
Pramendra Kumar, Gunawan Widjaja, Nikhil S. Patankar,
B. Suresh Kumar
19 A novel blockchain-based artificial
intelligence application for healthcare
automation
Abstract: The integration of blockchain and artificial intelligence (AI) technology can
facilitate a comprehensive healthcare automation platform that can leverage the
power of both AI and blockchain. This platform not only creates trust, enabling trans-
parency of patient data sharing and record keeping, but also allows AI and machine
learning algorithms to improve the efficiency and accuracy of healthcare technolo-
gies. It can provide healthcare data privacy, ease the access and sharing of medical
records and other confidential health information, and enable fast and secure medi-
cal imaging transfers.
19.1 Introduction
Artificial intelligence (AI)-driven analytics can be deployed to improve healthcare out-
comes, leading to more efficient data processing and diagnosis. In addition, the combi-
nation of blockchain and AI can enable a platform to track all health-related data,
including medical records, medical devices, laboratory data, medications, and other
treatments with full traceability [1]. Furthermore, blockchain technology can create
an immutable records system which fully stores medical and health data. This secure
data storage system can ensure the privacy, accuracy, and security of medical records
as well as protecting data against fraud, theft, and misuse.
19.1.1 Background
https://doi.org/10.1515/9783111323749-019
374 Pramendra Kumar et al.
of patients and give them timely diagnoses while allowing medical staff to focus on
providing quality healthcare instead of dealing with paperwork. Additionally, block-
chain technology further enhances the security and privacy of patient data as en-
crypted data is stored on distributed ledgers, making it safe from hackers.
The application of blockchain-based AI for healthcare automation is still in its
early development stages. However, its potential is already being realized in many
hospitals and clinics as it is helping to save time, money, and resources. In the near
future, these systems will play a major role in revolutionizing healthcare and making
it more efficient, secure, and affordable [4].
19.1.2 Objectives
ally, automation can provide better insights into population health by allowing for
more granular analysis and aggregation of patient data.
Healthcare automation is an important tool for improving the speed, accuracy, and
efficiency of medical processes, particularly those related to patient care. It can im-
prove patient outcomes by reducing errors, eliminating unnecessary manual pro-
cesses, increasing data accuracy, and providing easier access to patient records and
medical information [7]. Automation can also help healthcare professionals stay orga-
nized and better coordinate patient care, reducing wait times and improving overall
patient satisfaction. By increasing the overall efficiency of medical records and pro-
cesses, healthcare automation can save time and money, reduce human error, and
make the overall patient experience more efficient.
There are several challenges to healthcare automation, including the cost and com-
plexity of implementation, the need to maintain privacy and security, and the ability
to ensure that the technology is reliable. Automation requires a significant invest-
ment, as it involves complex hardware and software. Additionally, healthcare systems
must be extremely secure, as patient data and medical records are highly sensitive,
and any compromise of patient safety and confidentiality could lead to legal action
[8]. Finally, healthcare automation must be reliable and efficient, as medical person-
nel rely on the technology to provide medical care.
AI is a rapidly growing field that can be used to design, develop, and deploy solutions
to a variety of real-world problems. AI-based solutions offer a host of benefits, such as
automation, improved efficiency, enhanced accuracy, and reduced costs. AI has also
found applications in the healthcare industry in recent years [9].
AI-based tools, such as machine learning algorithms, can be used to automate vari-
ous processes, such as the assessment and diagnosis of medical conditions. AI-powered
technologies can also crunch large amounts of medical data in order to suggest more
accurate diagnoses and treatments [10]. Furthermore, AI can also be used to facilitate
medical research, assist with data management, and provide better patient care. Addi-
tionally, AI can be used to analyze the effectiveness of healthcare services and make
recommendations for improvement.
376 Pramendra Kumar et al.
In short, AI-based tools can be used to automate mundane tasks, improve the ac-
curacy of medical diagnoses, and identify inefficiencies in the delivery of healthcare
services. This technology has the potential to improve the quality of healthcare, re-
duce costs, and improve patient outcomes.
Blockchain technology is a type of DLT that enables the secure storage, transfer, and
management of data and digital assets, without the need for third-party intermediar-
ies. It is a decentralized, cryptographic system of potentially permanent records and
records that is available for anyone to view. The data is stored on a distributed net-
work of computers, and is secured by networks of computers running the consensus
algorithms, which verify the authenticity and accuracy of all the transactions within
the network [12]. This DLT has the potential to revolutionize the way financial trans-
actions are processed, and has become the foundation of a new economy. Blockchain
also provides a platform for smart contracts, enabling the automation of multiple pro-
cesses and operations within a company’s infrastructure.
Blockchain can be used in healthcare to securely store patient data, create an im-
mutable audit trail of all patient interactions, and streamline the exchange of medical
records between disparate systems. The technology can also be used to securely audit
claims and reimbursement procedures, manage financial trading and market settle-
ments, and manage clinical trials. Additionally, it can be used for population health
analytics, fraud detection, and biomedical research [14].
By cryptographically securing patient data on a decentralized ledger, blockchains
enable data to be securely shared between institutions and across different jurisdic-
tions, reducing administrative costs and improving access to care. This also enables
providers to have a better understanding of patient healthcare history, leading to
more targeted treatments and improved patient outcomes [15].
In addition, blockchain can be used to create digital currency networks that facili-
tate the transfer of value within healthcare systems. This could significantly improve
the efficiency of healthcare payments and create new opportunities for providers to
offer innovative health services. Ultimately, blockchain technology will help health-
care systems reduce costs, improve quality, and create an environment of trust.
rules and regulations that must be followed are slowing down the adoption of
blockchain technology in healthcare.
– Lack of knowledge and expertise: Even though blockchain technology is becom-
ing increasingly popular, lack of knowledge and expertise is a challenge. There is
a growing demand for experienced personnel to help implement and manage
healthcare applications using blockchain technology.
– Costs Involved: Even though blockchain technology can help reduce the cost of
some processes, the technology is pricey. Implementing and managing blockchain
technology requires the purchase of specialized hardware and expert services,
which can be costly.
These are just some of the AI techniques being used in healthcare automation today.
With continued advances in technology, AI solutions are becoming increasingly so-
phisticated and advanced, allowing healthcare professionals to provide better care
for their patients.
learning can be used to automate administrative tasks and make informed decisions
about patient care and treatment [16]. For example, machine learning algorithms can
be used to identify patterns in doctor–patient interactions, to create personalized
healthcare plans, and to detect anomalies in medical images. Machine learning also
enables healthcare providers to use vast amounts of data from patient records, lab
results, applications, and electronic health records (EHRs) to make decisions about di-
agnosis and treatment. By leveraging machine learning, healthcare providers can
make more accurate decisions at a faster pace and reduce costs by streamlining pro-
cesses. Additionally, machine learning can be used to create predictive models to an-
ticipate the outcome of treatments and give insights on patient health [17].
Blockchain-based data sharing and security is a secure way to send and receive data
leveraging the power of blockchain technology. It creates an immutable, distributed
ledger of transactions that all users can trust and verify. With this technology all par-
ticipants of the network agree on a shared version of the data, eliminating the need
for a central authority or a centralized storage [27]. Data is encrypted and stored in a
distributed ledger, where it is shared and updated in real time between all partici-
pants in the network. This makes it much harder for data to be compromised or al-
tered. Additionally, this type of data sharing ensures better data privacy by using
public-key cryptography to encrypt data before it’s sent. With blockchain technology,
data can be shared securely without relying on a single source of truth.
A smart contract for healthcare is an automated and digital agreement between par-
ties, designed to facilitate the execution of healthcare-related transactions. The pur-
pose of these contracts is to enable improved efficiency, cost-effectiveness, accuracy,
trust, and security in the healthcare industry [28].
Smart contracts can be used for a variety of healthcare activities, such as pay-
ment processing, insurance claims, and record keeping. By automating these pro-
cesses, they reduce the need for face-to-face transactions and eliminate the potential
for paper-based errors. By securely storing healthcare data, smart contracts help to
prevent data loss. Furthermore, they allow for instant payment processing, secure ar-
chiving of patient records, and improved accuracy through automated healthcare an-
alytics [29].
Smart contracts also provide enhanced security and privacy measures that are not
possible using traditional healthcare models. Through encryption of sensitive data, im-
proved authentication protocols, and improved access control, smart contracts mini-
mize the risk of unauthorized access to confidential information. Furthermore, these
contracts are designed to provide improved accountability and transparency in the
healthcare industry, allowing all stakeholders to feel secure in the knowledge that their
data is safe and secure [30].
fessionals – such as health information managers – who are responsible for the accu-
rate collection, maintenance, storage, security, and sharing of patient data. These pro-
fessionals must ensure that all data is accurately entered and updated, and must also
adhere to all applicable HIPAA regulations. They also need to provide additional sup-
port to providers who are using EHRs, such as helping them learn how to enter data
and interpret reports.
The medical supply chain tracking is a process of tracking the movement of medical
products and services from the source of supply to the point of care. It involves the
tracking of both physical goods and digital transactions. This process helps to ensure
that the right product is delivered to the right patient at the right time, and that the
inventory of each medical supplier is kept up to date.
The medical supply chain tracking process starts at the supplier, tracking the de-
livery of products from the vendor to the doctor’s office, hospital, or other care pro-
vider. The tracking also includes the dispensing of the product or service, the receipt
of payment, and the disposal or return of the product or service. Depending on the
type of product or service being tracked, the tracking system may include additional
information such as patient prepayments, orders, credit, or authorization codes.
The tracking system can provide useful data for analyzing performance, identify-
ing trends, and improving patient care. This data can be used to help identify and mit-
igate supply chain risks, such as stock levels and delivery times, as well as to ensure
compliance with regulatory requirements and government standards. In addition,
tracking systems can be used to automate functions such as inventory tracking and
ordering processes, thereby reducing paperwork and improving efficiency.
References
[1] Allen, B., Dreyer, K., Stibolt, R., Jr, Agarwal, S., Coombs, L., Treml, C., . . . & Wald, C. (2021).
Evaluation and real-world performance monitoring of artificial intelligence models in clinical
practice: Try it, buy it, check it. Journal of the American College of Radiology, 18(11), 1489–1496.
[2] Tabbussum, R., & Dar, A. Q. (2021). Performance evaluation of artificial intelligence paradigms –
artificial neural networks, fuzzy logic, and adaptive neuro-fuzzy inference system for flood
prediction. Environmental Science and Pollution Research, 28(20), 25265–25282.
[3] Hicks, S. A., Strümke, I., Thambawita, V., Hammou, M., Riegler, M. A., Halvorsen, P., & Parasa, S. (2022).
On evaluation metrics for medical applications of artificial intelligence. Scientific Reports, 12(1), 5979.
[4] Li, L., Qin, L., Xu, Z., Yin, Y., Wang, X., Kong, B., . . . & Xia, J. (2020). Using artificial intelligence to
detect COVID-19 and community-acquired pneumonia based on pulmonary CT: Evaluation of the
diagnostic accuracy. Radiology, 296(2), E65–E71.
384 Pramendra Kumar et al.
[5] Dumitrascu, O., Dumitrascu, M., & Dobrotǎ, D. (2020). Performance evaluation for a sustainable
supply chain management system in the automotive industry using artificial intelligence. Processes,
8(11), 1384.
[6] Ghahramani, M., Qiao, Y., Zhou, M. C., O’Hagan, A., & Sweeney, J. (2020). AI-based modeling and
data-driven evaluation for smart manufacturing processes. IEEE/CAA Journal of Automatica Sinica,
7(4), 1026–1037.
[7] Khanagar, S. B., Al-Ehaideb, A., Maganur, P. C., Vishwanathaiah, S., Patil, S., Baeshen, H. A., . . . &
Bhandi, S. (2021). Developments, application, and performance of artificial intelligence in dentistry –
A systematic review. Journal of Dental Sciences, 16(1), 508–522.
[8] Wu, E., Wu, K., Daneshjou, R., Ouyang, D., Ho, D. E., & Zou, J. (2021). How medical AI devices are
evaluated: Limitations and recommendations from an analysis of FDA approvals. Nature Medicine,
27(4), 582–584.
[9] Amin, M. M., Cambria, E., & Schuller, B. W. (2023). Will affective computing emerge from foundation
models and general AI? A first evaluation on chatgpt. In 2023 International Conference on
computation and Language (pp. 761–768). IEEE, Singapore.
[10] Forghani, R., Savadjiev, P., Chatterjee, A., Muthukrishnan, N., Reinhold, C., & Forghani, B. (2019).
Radiomics and artificial intelligence for biomarker and prediction model development in oncology.
Computational and Structural Biotechnology Journal, 17, 995.
[11] Tobler, P., Cyriac, J., Kovacs, B. K., Hofmann, V., Sexauer, R., Paciolla, F., . . . & Hirschmann, A. (2021).
AI-based detection and classification of distal radius fractures using low-effort data labeling:
Evaluation of applicability and effect of training set size. European Radiology, 31, 6816–6824.
[12] AI-Atroshi, C., Rene Beulah, J., Singamaneni, K. K., Pretty Diana Cyril, C., Neelakandan, S., &
Velmurugan, S. (2022). Automated speech based evaluation of mild cognitive impairment and
Alzheimer’s disease detection using with deep belief network model. International Journal of
Healthcare Management, 2022, 1–11.
[13] Gunning, D., & Aha, D. (2019). DARPA’s explainable artificial intelligence (XAI) program. AI Magazine,
40(2), 44–58.
[14] Zou, Y., Zheng, C., Alzahrani, A. M., Ahmad, W., Ahmad, A., Mohamed, A. M., . . . & Elattar, S. (2022).
Evaluation of artificial intelligence methods to estimate the compressive strength of geopolymers.
Gels, 8(5), 271.
[15] Yadav, S. P., & Yadav, S. (2020). Image fusion using hybrid methods in multimodality medical
images. Medical and Biological Engineering and Computing, 58, 669–687.
[16] Irfan, D., Tang, X., Narayan, V., Mall, P. K., Srivastava, S., & Saravanan, V. (2022). Prediction of quality
food sale in mart using the AI-based TOR method. Journal of Food Quality, 2022, 1–13.
[17] Yuvaraj, N., Srihari, K., Dhiman, G., Somasundaram, K., Sharma, A., Rajeskannan, S. M. G. S. M. A., . . .
& Masud, M. (2021). Nature-inspired-based approach for automated cyberbullying classification on
multimedia social networking. Mathematical Problems in Engineering, 2021, 1–12.
[18] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). Energy‐efficient resource allocation
model for device‐to‐device communication in 5G wireless personal area networks. International
Journal of Communication Systems, 2023, 1–14.
[19] Yadav, S. P., Agrawal, K. K., Bhati, B. S., Al-Turjman, F., & Mostarda, L. (2022). Blockchain-based
cryptocurrency regulation: An overview. Computational Economics, 59(4), 1659–1675.
[20] Rajalakshmi, M., Saravanan, V., Arunprasad, V., Romero, C. T., Khalaf, O. I., & Karthik, C. (2022).
Machine learning for modeling and control of industrial clarifier process. Intelligent Automation
and Soft Computing, 32(1), 339–359.
[21] Manikandan, V., & Yuvaraj, N. (2019, May). Advanced expert system using particle swarm
optimization based adaptive network based fuzzy inference system to diagnose the physical
constitution of human body. In Emerging Technologies in Computer Engineering: Microservices in
19 A novel blockchain-based artificial intelligence application 385
Big Data Analytics: Second International Conference, ICETCE 2019, Jaipur, India, February 1–2, 2019,
Revised Selected Papers (Vol. 985, p. 349). New York: Springer.
[22] Sangeetha, R., Logeshwaran, J., Rocher, J., & Lloret, J. An improved agro deep learning model for
detection of Panama wilts disease in banana leaves. AgriEngineering. 2023; 5(2):660–679
[23] Yadav, S. P., Zaidi, S., Mishra, A., & Yadav, V. (2022). Survey on machine learning in speech emotion
recognition and vision systems using a recurrent neural network (RNN). Archives of Computational
Methods in Engineering, 29(3), 1753–1770.
[24] Saravanan, V., & Neeraja, A. (2013, January). Security issues in computer networks and
stegnography. In 2013 7th International Conference on Intelligent Systems and Control (ISCO)
(pp. 363–366). IEEE.
[25] Gobinathan, B., Mukunthan, M. A., Surendran, S., Somasundaram, K., Moeed, S. A., Niranjan, P., . . .
& Sundramurthy, V. P. (2021). A novel method to solve real time security issues in software industry
using advanced cryptographic techniques. Scientific Programming, 2021, 1–9.
[26] Logeshwaran, J., Kiruthiga, T., & Lloret, J., (2023). A novel architecture of intelligent decision model
for efficient resource allocation in 5G broadband communication networks. ICTACT Journal on Soft
Computing, 13 (3), 2986–2994
[27] Yadav, S. P., Mahato, D. P., & Linh, N. T. D. (Eds.). (2020). Distributed artificial intelligence: A modern
approach. Boca Raton: CRC Press.
[28] Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J., . . . &
Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey
wolf optimization in wireless sensor networks. Security and Communication Networks, 2022.
[29] Natarajan, Y., Kannan, S., & Mohanty, S. N. (2021). Survey of various statistical numerical and
machine learning ontological models on infectious disease ontology. Data Analytics in
Bioinformatics: A Machine Learning Perspective, 14(1), 431–442.
[30] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). L-RUBI: An efficient load-based
resource utilization algorithm for bi-partite scatternet in wireless personal area networks.
International Journal of Communication Systems, 36(4), 1–17.
[31] Wan, Y., Tong, Y., Liu, Y., Huang, Y., Yao, G., Chen, D. Q., & Liu, B. (2022). Evaluation of the
combination of artificial intelligence and radiologist assessments to interpret malignant
architectural distortion on mammography. Frontiers in Oncology, 12, 1739.
[32] Hosamo, H. H., Nielsen, H. K., Kraniotis, D., Svennevig, P. R., & Svidt, K. (2023). Digital Twin
framework for automated fault source detection and prediction for comfort performance evaluation
of existing non-residential Norwegian buildings. Energy and Buildings, 281, 112732.
[33] Lee, D. H., Woo, S. E., Jung, M. W., & Heo, T. Y. (2022). Evaluation of odor prediction model
performance and variable importance according to various missing imputation methods. Applied
Sciences, 12(6), 2826.
[34] Markus, A. F., Kors, J. A., & Rijnbeek, P. R. (2021). The role of explainability in creating trustworthy
artificial intelligence for health care: A comprehensive survey of the terminology, design choices,
and evaluation strategies. Journal of Biomedical Informatics, 113, 103655.
[35] Orhan, K., Bayrakdar, I. S., Ezhov, M., Kravtsov, A., & Özyürek, T. A. H. A. (2020). Evaluation of
artificial intelligence for detecting periapical pathosis on cone‐beam computed tomography scans.
International Endodontic Journal, 53(5), 680–689.
[36] Schwartz, R., Dodge, J., Smith, N. A., & Etzioni, O. (2020). Green ai. Communications of the ACM,
63(12), 54–63.
Soumya Mishra, Priyanka Sharma, Bhoopathy Varadharajan,
Bhola Khan
20 Enhancing industrial efficiency with
AI-enabled blockchain-based solutions
Abstract: The ever-evolving technologies are constantly making the industrial pro-
cesses more efficient and providing new opportunities for cost saving and enhanced
operations. Artificial intelligence (AI)-enabled blockchain-based solutions are making
significant contributions to the enhancement of industrial efficiency. AI-enabled
blockchain-based solutions enable the secure and frictionless exchange of digital as-
sets and data among members of the chain while enabling trustless interaction be-
tween them. These solutions enable faster and error-free transactions with enhanced
scalability, reliability, traceability, and privacy.
20.1 Introduction
Artificial intelligence (AI)-enabled blockchain-based solutions can enable automatic
tracking of products through the supply chain, which helps reduce time and cost
while ensuring the accuracy of data in real time. These solutions also make data shar-
ing more secure, as they enable the secure storage of transaction information which
can be utilized for data analytics in the future [1]. Additionally, AI-enabled block-
chain-based solutions can be utilized to identify and analyze potential fraudulent ac-
tivities within the chain. This helps in reducing the risks associated with fraudulent
activities, and increases safety and security. AI-enabled blockchain-based solutions
are valuable tools that are making industrial processes more efficient.
https://doi.org/10.1515/9783111323749-020
388 Soumya Mishra et al.
smarter and more efficient industrial operations and processes, leading to improved
productivity and cost savings for businesses.
Industrial efficiency refers to the use of systems, machines, and technologies that re-
duce inflation and improve outputs. This helps companies cut costs, become more ef-
ficient, and maximize productivity [6].
However, like any other business process, industrial efficiency can be quite chal-
lenging. Some of the challenges include:
20 Enhancing industrial efficiency with AI-enabled blockchain-based solutions 389
IDs and ensure better supply chain traceability through smart contracts, streamlining
the process of sharing and managing data in the automotive sector [9].
In the logistics industry, blockchain technology can be used to create efficient, au-
tomated systems for verifying asset location, tracking, and transfer of ownership. This
technology can be used to facilitate secure transactions, such as payments, with the
assurance that digital assets can be transferred safely and securely. Additionally,
blockchain can also enable faster and more secure delivery of goods and services, as
well as improve the tracking and verification of shipments.
Blockchain technology also has applications in healthcare, such as with its ability
to securely store and manage patient data. Through blockchain, data can be shared
with multiple parties safely and securely, while ensuring regulatory compliance obli-
gations are met. This technology can also be used to create digital identities for pa-
tients, allowing them to manage their healthcare records securely and efficiently [10].
The blockchain technology is revolutionizing many different industries, from fi-
nance to automotive to logistics and healthcare. It offers a secure, reliable, and effi-
cient platform for managing digital assets, creating digital identities, and tracking
data. As the technology continues to evolve and gain traction, the potential for its use
in industrial applications is growing.
chain, industrial organizations can work more efficiently and smarter than ever
before.
Challenges:
– Complexity: Blockchain technology is relatively new and still evolving, which can
make it difficult for some organizations to implement. Moreover, there are still
issues concerning scalability, interoperability, and data privacy that must be
addressed.
– Cost: Implementing blockchain technology can require significant investment in
infrastructure and resources. This could be a barrier for some companies that are
not willing or able to invest in this technology.
– Regulation: As blockchain technology is still relatively new, there is a lack of regu-
latory framework in many jurisdictions. This can hamper the development and
adoption of blockchain technology, as organizations may be unwilling to take
risks without proper guidance.
for production, design, maintenance, and other processes and to detect defects. AI can
also be used to monitor the operational environment, provide recommendations for
process improvement, and automate tasks to reduce human labor. AI applications can
help industrial companies save time, money, and resources in production, while keep-
ing quality at a high level. This can enable companies to be more agile and adaptive
and respond more quickly to changing needs and opportunities.
Machine learning in industrial efficiency includes the use of AI and machine learning
algorithms to automate the processes, tasks, workflows, and operational decisions
that make up traditional industries. By offering both automated decision-making ca-
pabilities and instances of valuable insights on how to improve operations, machine
learning helps companies operate more efficiently and effectively [14].
Industrial applications, such as logistics and supply chain management, quality
assurance, maintenance scheduling, production optimization, safety, and predictive
analytics, can be improved through the application of machine learning.
In addition to algorithmic decision-making, machine learning can also be used to
pre-process, analyze, and interpret data from sensors, machines, and other devices
throughout a facility. This data can then be used to create digital twins of real assets,
create probabilistic models about uncertain future events and conditions, and provide
predictive insights about potential maintenance needs. For example, machine learn-
ing can be used to forecast future asset performance, alert operators of potential
maintenance needs before they occur, and provide advice on the best course of action
for the situation [15].
It can also be used to improve efficiency across multiple machines by suggesting
adjustments to control parameters that achieve maximum yields or minimum energy
consumption, for example. Ultimately, machine learning can help organizations re-
duce costs, reduce downtime, increase yield, optimize production, and increase over-
all safety.
Robots are the most common form of industrial automation. They are computer-
controlled, programmable machines that can be programmed to perform a wide
range of tasks. These tasks can range from highly complicated tasks such as welding
and painting to simpler tasks such as picking and packing. Robots are often used in
manufacturing to perform repetitive assembly tasks that would otherwise require a
large number of manpower.
Automation systems are also used in industrial settings. These systems are net-
works of machines and software that can be programmed to perform multiple tasks.
Automation systems can be used for data processing, material movement, quality con-
trol, and production scheduling. Automation systems can also be used for controlling
complex systems such as production lines and assembly operations [17].
The robotics and automation are becoming increasingly important in industrial
settings. They provide companies with the ability to increase production rates, reduce
labor costs, and maintain quality standards. With the right implementation, robotics
and automation can greatly improve the efficiency and effectiveness of a manufactur-
ing operation.
up-to-date. This would help in reducing fraud and ensuring the efficient utilization of
resources. Furthermore, blockchain can also be employed to monitor energy usage
and enforce energy consumption rules which have been sanctioned by the govern-
ment authorities [19]. In addition, it can also provide an immutable ledger to record
the transactions that take place between different entities in the supply chain. This
would allow for greater accountability, appropriate pricing mechanisms, and secure
data sharing.
making, and process automation [22]. Autonomous systems are self-governing and
can operate without any human intervention. They are designed to ensure accuracy,
reliability, and speed of the contractual process. AI-enabled smart contracts and au-
tonomous systems are revolutionizing the way we manage and execute contracts and
transactions. They provide a secure environment and help reduce time, costs, and op-
erational risks.
Data integrity and security in industrial processes is a set of processes and technolo-
gies that are needed to ensure the safe and secure operation of industrial process sys-
tems. By implementing data and system security measures, organizations can ensure
that the data that is collected and stored is not compromised or inappropriately ac-
cessed. This is done by employing techniques such as encryption, firewalls, user au-
thentication, and access control [23]. This data security layer helps protect against
unauthorized access and malicious attacks. Keeping data integrity also means that
data is free from unauthorized use, alteration, and destruction. Data integrity is main-
tained through controls such as user authentication, data encryption, data backup,
and storage policies. Additionally, companies must have a continuous process in place
to keep data secure and to monitor the security of their data systems.
Supply chain optimization is the process of making a supply chain more efficient by
continuously analyzing and improving the activities and processes used to deliver
396 Soumya Mishra et al.
Energy management and sustainability refer to the practice and process of using energy
in a way that can improve overall energy efficiency. This includes the use of energy
from renewable sources, the use of energy efficient technologies, and the creation of
sustainable energy plans. Sustainability is an approach to energy management that
seeks to ensure that energy resources are used more responsibly over time, while mini-
mizing environmental impacts. The goal of energy management and sustainability is to
reduce energy consumption and improve the efficiency of energy use, which in turn
reduces resource consumption and emissions. Energy efficiency saves money, reduces
the need for electricity or fuel supply, and reduces greenhouse gas emissions [26]. Sus-
tainability initiatives such as energy audits, benchmarking, and implementing smart en-
ergy strategies can improve energy utilization, enhance system functionality, and
reduce operating costs.
Quality control (QC) is the process of ensuring that a product or service meets a cer-
tain set of quality standards. QC includes inspecting, testing, and validating a product
or delivery service to meet customer requirements and expectations. QC is used to
identify defects in products or services, and to guarantee that they conform to estab-
lished standards.
Defect detection is the process of identifying defects or errors in a product or ser-
vice and taking corrective action. Defects can be physical defects or deficiencies in the
functionality of the product or service. Defect detection can be done through manual
and automated methods. Manual inspection is typically used for low production rates,
while automated testing is used when production rates are high.
20 Enhancing industrial efficiency with AI-enabled blockchain-based solutions 397
References
[1] Haenlein, M., & Kaplan, A. (2019). A brief history of artificial intelligence: On the past, present, and
future of artificial intelligence. California Management Review, 61(4), 5–14.
[2] Zhang, C., & Lu, Y. (2021). Study on artificial intelligence: The state of the art and future prospects.
Journal of Industrial Information Integration, 23, 100224.
[3] Schwalbe, N., & Wahl, B. (2020). Artificial intelligence and the future of global health. Lancet I,
395(10236), 1579–1586.
[4] Schiff, D. (2021). Out of the laboratory and into the classroom: The future of artificial intelligence in
education. AI and Society, 36(1), 331–348.
[5] Belfield, H. (2020, February). Activism by the AI community: Analysing recent achievements and
future prospects. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (pp.15–21).
[6] Biswas, S., Carson, B., Chung, V., Singh, S., & Thomas, R. (2020). AI-bank of the future: Can banks
meet the AI challenge. New York: McKinsey & Company.
[7] Cave, S. (2020, February). The problem with intelligence: Its value-laden history and the future of AI.
In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (pp.29–35).
[8] Sur, J., Bose, S., Khan, F., Dewangan, D., Sawriya, E., & Roul, A. (2020). Knowledge, attitudes, and
perceptions regarding the future of artificial intelligence in oral radiology in india: A survey.
Imaging Science in Dentistry, 50(3), 193.
[9] Aggarwal, K., Mijwil, M. M., Al-Mistarehi, A. H., Alomari, S., Gök, M., Alaabdin, A. M. Z., & Abdulrhman,
S. H. (2022). Has the future started? The current growth of artificial intelligence, machine learning, and
deep learning. Iraqi Journal for Computer Science and Mathematics, 3(1), 115–123.
[10] Sangeetha, R., Logeshwaran, J., Rocher, J., & Lloret, J. (2023). An improved agro deep learning model
for detection of Panama wilts disease in banana leaves. AgriEngineering, 5(2), 660–679.
[11] Singh, P. K., Chauhan, S. S., Sharma, A., Prakash, S., & Singh, Y. (2023). Prediction of higher heating
values based on imminent analysis by using regression analysis and artificial neural network for
bioenergy resources. Proceedings of the Institution of Mechanical Engineers, Part E: Journal of
Process Mechanical Engineering, 09544089231175046.
[12] Chauhan, S. S., Yadav, S. P., Awashthi, S., & Naruka, M. S. (2023). Sentimental analysis using cloud
dictionary and machine learning approach. In: Kaushal Kishor, Neetesh Saxena, Dilkeshwar Pandey
(Eds.), Cloud-based Intelligent Informative Engineering for Society 5.0 (pp. 157–169). Boca Raton:
Chapman and Hall/CRC.
[13] Awasthi, S., Naruka, M. S., Yadav, S. P., & de Albuquerque, V. H. C. (Eds.). (2023). AI and IoT-based
intelligent health care & sanitation. United Arab Emirates: Bentham Science Publishers.
[14] Yadav, S. P., Zaidi, S., Nascimento, C. D. S., De albuquerque, V. H. C., & Chauhan, S. S. (2023, January).
Analysis and design of automatically generating for GPS based moving object tracking system. In 2023
International Conference on Artificial Intelligence and Smart Communication (AISC) (pp. 1–5). IEEE.
[15] Saravanan, V., Sumathi, A., Shanthana, S., & Rizvana, M. (2013). Dual mode mpeg steganography
scheme for mobile and fixed devices. International Journal of Engineering Research and
Development, 6, 23–27.
400 Soumya Mishra et al.
[16] Cihon, P., Maas, M. M., & Kemp, L. (2020). Fragmentation and the future: Investigating architectures
for international AI governance. Global Policy, 11(5), 545–556.
[17] Stahl, B. C. (2021). Artificial intelligence for a better future: An ecosystem perspective on the ethics
of AI and emerging digital technologies (p. 124). New York: Springer Nature.
[18] Allam, Z., Dey, G., & Jones, D. S. (2020). Artificial intelligence (AI) provided early detection of the
coronavirus (COVID-19) in China and will influence future Urban health policy internationally. Ai, 1
(2), 156–165.
[19] Saravanan, V., & Chandrasekar, C. (2012). Qos-continuous live media streaming in mobile environment
using vbr and edge network. International Journal of Computer Applications, 53(6), 1–12.
[20] Lakshminarayanan, R., & Yuvaraj, N. (2020). Design and analysis of an improved deep learning
algorithm on classification of intervertebral discs. International Journal of Advanced Science and
Technology, 29, 4019–4026.
[21] Gupta, S., Saravanan, V., Choudhury, A., Alqahtani, A., Abonazel, M. R., & Babu, K. S. (2022).
Supervised computer-aided diagnosis (CAD) methods for classifying alzheimer’s disease-based
neurodegenerative disorders. Computational and Mathematical Methods in Medicine, 2022, 1–9.
[22] Indhumathi, R., Amuthabala, K., Kiruthiga, G., Yuvaraj, N., & Pandey, A. (2023). Design of task
scheduling and fault tolerance mechanism based on GWO Algorithm for attaining better QoS in
cloud system. Wireless Personal Communications, 128(4), 2811–2829.
[23] Zhang, K., & Aslan, A. B. (2021). AI technologies for education: Recent research & future directions.
Computers and Education: Artificial Intelligence, 2, 100025.
[24] Vaishya, R., Javaid, M., Khan, I. H., & Haleem, A. (2020). Artificial intelligence (AI) applications for COVID-
19 pandemic. Diabetes and Metabolic Syndrome: Clinical Research and Reviews, 14(4), 337–339.
[25] Aruna R, D., Surendran S, D., Yuvaraj N, D., & Debtera, B. (2022). An enhancement on convolutional
artificial intelligent based diagnosis for skin disease using nanotechnology sensors. Computational
Intelligence and Neuroscience, 2022, 1–11.
[26] Berente, N., Gu, B., Recker, J., & Santhanam, R. (2021). Managing artificial intelligence. MIS
Quarterly, 45(3), 1433–1450.
[27] Yuvaraj, N., Praghash, K., & Karthikeyan, T. (2022). Privacy preservation of the user data and
properly balancing between privacy and utility. International Journal of Business Intelligence and
Data Mining, 20(4), 394–411.
[28] Ramesh, G., Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). Prediction of energy production level
in large PV plants through AUTO-Encoder Based Neural-Network (AUTO-NN) with restricted
Boltzmann feature extraction. Future Internet, 15(2), 46.
[29] Saravanan, V., Sreelatha, P., Atyam, N. R., Madiajagan, M., Saravanan, D., & Sultana, H. P. (2023).
Design of deep learning model for radio resource allocation in 5G for massive iot device.
Sustainable Energy Technologies and Assessments, 56, 103054.
[30] Borges, A. F., Laurindo, F. J., Spínola, M. M., Gonçalves, R. F., & Mattos, C. A. (2021). The strategic use
of artificial intelligence in the digital era: Systematic literature review and future research directions.
International Journal of Information Management, 57, 102225.
[31] Lu, Y. (2019). Artificial intelligence: A survey on evolution, models, applications and future trends.
Journal of Management Analytics, 6(1), 1–29.
[32] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). L-RUBI: An efficient load-based
resource utilization algorithm for bi-partite scatternet in wireless personal area networks.
International Journal of Communication Systems, 36(4), 1–17.
[33] Fouad, H., Hassanein, A. S., Soliman, A. M., & Al-Feel, H. (2020). Analyzing patient health information
based on IoT sensor with AI for improving patient assistance in the future direction. Measurement,
159, 107757.
Index
abnormalities 215 anonymization 196, 317
account 277, 279–280, 289, 292, 294 anticipate 254
accountability 189–190, 196–198, 203–206, 208 antiquated 39
accountable 190, 192, 194, 198 appearances 40
accuracy 1–2, 6, 10–11, 14, 93, 104–105, 107, applications 2, 5–6, 7, 9–12, 15, 17, 154–155, 161,
125–141, 145–149, 153–155, 157, 159–160, 233, 235, 287–291, 293, 297, 302–304, 331,
162, 168–170, 173–183, 186, 253, 269–274, 276, 333, 335–339
278, 281 approach 167, 170, 173–174, 180–181, 183, 185
accurate 23–24, 26, 33, 35–36, 38, 41, 49, 55–56, approaches 109, 113, 115–118
58, 62, 67–69, 71, 73–75, 77–80, 216–217, 220, Apriori 87, 95, 106
226–227 Apriori algorithm 87, 95, 106
acquisition 224 architectures 5, 299
ad hoc network 333 artificial 1–2, 3, 4, 9, 13–14, 213, 215–217, 219,
adaptable 90 224–226, 228, 233, 235–236, 238, 241,
addressing 202, 206, 208 243–246, 251, 253–258, 261–262, 269, 287,
adjustable 174 291, 361–362, 365, 367–368
administrators 339 artificial intelligence 1, 3–4, 5, 6, 7, 8, 9, 10, 11,
adversarial training 121 13–18, 291–292, 294, 297, 345, 348, 350, 352,
advertising 246 354–355
affordable 374 assessments 129
agencies 137 association 24, 345
agents 118–121, 271, 274–275 assumptions 152, 154, 161, 200, 205
aggregating 153, 168, 170, 177 asymmetry 99
agile 282 audiences 315, 321
agreement 128–129 audit 204
AI 21–23, 27–28, 30, 41, 43–49, 345–346, 350, augmentation 190, 199
352–357, 373–376, 378–382, 387, 391–395, augmented 328
397–399 authentication 310, 316–318, 325
AI-driven 251–254, 256–259, 261–262, 264, 266 authorized 317
algorithm 126–127, 131, 134–135 automate 233–241, 246, 287–288, 291, 293,
algorithms 1, 3–4, 5, 7, 9–13, 16, 21–26, 29–32, 34, 295–297, 302–305
36–39, 42, 46, 53, 61–65, 68, 70–76, 79, 81–83, automated 2–3, 6, 9, 13, 361–363, 367, 370
167–168, 171–173, 175, 177–178, 180–185, automation 22, 41, 43, 45, 192, 201, 233–234,
189–195, 200, 208, 253–255, 257–261, 238–239, 242, 246, 269–274, 276–278,
263–264, 266, 288, 291–296, 302–303, 305, 280–283, 373–376, 378–379
332, 334–340 automotive 389–390
allocations 218 autonomous 23, 37–38, 40, 42, 44, 214, 219
alternative 109 autonomous systems 7
Amazon 311 autonomy 302
ambiguities 347 average 130, 133–134, 168–169, 171
ambiguity 364
analysis 345–357, 361–363, 365, 367–370 bagging 153, 167, 169–171, 176–181, 184
analytics 310, 315, 317–319, 323, 326 banks 18
analyze 345–346, 348, 350–352, 354, 357 Bayesian 235
annealing 155, 292 behaviors 297, 303
anomalies 21, 23, 36, 40, 42, 47 beneficial 201, 203, 207
anomaly 24, 173, 175, 233, 235 benefit 241, 288
https://doi.org/10.1515/9783111323749-021
402 Index
e-commerce 2–3, 32, 95, 97, 309–311, 314, 316–317, facial recognition 3–4, 7, 15, 38, 40
324–328 factors 318, 320, 324
economic 200, 207 fairness 189–193, 196–198, 203, 206
economy 13 fallibility 199
education 288, 303 false negatives 126
404 Index