Professional Documents
Culture Documents
ETI Microproject 1
ETI Microproject 1
On
Of
Manasvi Ambavale
Pawan Pawar
Raj Fadtare
Chinmay Sawant
(2023-2024)
1
MAHARASHTRA STATE BOARD OF TECHNICAL
EDUCATION
Certificate
The is to certify that Mr./Ms. of 6th semester of Diploma in Information
Technology of Institute, Kala Vidya Mandir Institute of Technology
(Code:0571) has completed the Micro Project satisfactorily in Subject:
ETI for the academic year (2023-2024) as prescribed in the curriculum.
2. Introduction 5.
3. Machine learning 6.
5. Deep learning 8.
6. Chat Gpt 9.
7. Example’s 10-11.
8. References 12.
9. Conclusion 13.
3
Aim
The aim of a case study about artificial intelligence (AI) can vary based on the specific
context, industry, and objectives. Here are several potential aims for a case study on AI:
Assessing Future Trends: Predict future trends in AI adoption and innovation based on
current case studies and industry developments. Discuss potential opportunities and
challenges that may arise as AI continues to evolve.
4
Introduction
According to some, artificial intelligence is the most promising development for the future.
From curing cancer to resolving the global hunger crisis, artificial intelligence is being
presented as the solution to all of our problems. Others, however, regard it as a threat
artificial intelligence may potentially give rise to unemployment and inequality, and could
even jeopardize the continued existence of humankind. As the technology entrepreneur
Elon Musk put it: “The benign scenario is that artificial intelligence can do any job that
humans do – but better.” Deloitte has positioned itself on the optimistic side of that
spectrum. “We believe that artificial intelligence will be extremely helpful to us and to our
clients”, says Richard Roovers, a partner at Deloitte Netherlands and Innovation Lead
Transformational Solutions North-West Europe. Artificial intelligence will enable us to
solve problems that humans are unable, or hardly capable, of solving, explains Richard.
“Artificial intelligence is capable of processing massive quantities of data and has the
ability to discover patterns that even the smartest mathematicians are unable to find. That
in itself opens up a large number of new possibilities.” Those new possibilities are what
this book is about. The case studies provide an overview of the ways in which Deloitte is
working to develop applications incorporating artificial intelligence – both internally and
for use with clients. The applications are diverse, make use of different technologies and
can be found in a diverse range of industries. This shows that aside from all of the
predictions for the future, artificial intelligence has already been a reality in the business
sector for some time and forms a resource that could possibly provide your company with
a decisive lead. The concept of AI dates back several decades, but recent breakthroughs in
computational power, big data, and algorithmic sophistication have propelled its rapid
evolution. Today, AI applications permeate nearly every aspect of our lives, from virtual
assistants on our smartphones to predictive algorithms powering financial markets, from
personalized recommendations on streaming platforms to autonomous vehicles navigating
city streets. One of the key drivers behind the proliferation of AI is its ability to analyse
vast amounts of data with remarkable speed and accuracy, uncovering insights and patterns
that would be virtually impossible for humans to discern. This data-driven approach has
unlocked new opportunities for businesses to enhance decision-making, optimize
processes, and create personalized experiences for customers. However, the widespread
adoption of AI also raises important ethical, societal, and economic considerations.
Concerns about data privacy, algorithmic bias, job displacement, and the concentration of
power in the hands of a few tech giants have sparked debates about the responsible
development and deployment of AI technologies. In this age of rapid technological change,
understanding the capabilities, limitations, and implications of AI is more important than
ever.
5
Machine Learning
Machine learning is the study of programs that can improve their performance on a given
task automatically. It has been a part of AI from the beginning.
There are several kinds of machine learning. Unsupervised learning analyses a stream of
data and finds patterns and makes predictions without any other guidance. Supervised
learning requires a human to label the input data first, and comes in two main varieties:
classification (where the program must learn to predict what category the input belongs in)
and regression (where the program must deduce a numeric function based on numeric
input). In reinforcement learning the agent is rewarded for good responses and punished
for bad ones. The agent learns to choose responses that are classified as
"good". Transfer
learning is when the
knowledge gained from one
problem is applied to a new
problem.[47] Deep learning is
a type of machine learning
that runs inputs through
biologically inspired
artificial neural
networks for all of these
types of learning.[48]
Computational learning
theory can assess learners
by computational
complexity, by sample
complexity (how much data
is required), or by other notions of optimization.
6
Natural language processing
Natural language processing (NLP) allows programs to read, write and communicate in
human languages such as English. Specific problems include speech recognition, speech,
synthesis, machine translation, information extraction, information retrieval and question
answering.[51] Early work, based on Noam Chomsky's generative grammar and semantic
networks, had difficulty with word-sense disambiguation[f] unless restricted to small
domains called "micro-worlds" (due to the common sense knowledge problem). Margaret
Master man believed that it was meaning, and not grammar that was the key to
understanding languages, and that thesauri and not dictionaries should be the basis of
computational language structure. Modern deep learning techniques for NLP include
word embedding (representing words, typically as vectors encoding their
meaning), transformers (a deep learning architecture using an attention ), and others. In
2019, generative pre-trained transformer (or "GPT") language models began to generate
coherent text, and by 2023 these
models were able to get human-
level scores on the bar
exam, SAT test, GRE test, and
many other real-world
application. Natural language
processing (NLP) is a subfield
of Artificial Intelligence (AI).
This is a widely
used technology for
personal assistants that
are used in
various business fields/areas. This technology works on the speech provided by the user
breaks it down for proper understanding and processes it accordingly. This is a very recent
and effective approach due to which it has a really high demand in today’s market. Natural
Language Processing is an upcoming field where already many transitions such as
compatibility with smart devices, and interactive talks with a human have been made
possible. Knowledge representation, logical reasoning, and constraint satisfaction were the
emphasis of AI applications in NLP. Here first it was applied to semantics and later to
grammar. In the last decade, a significant change in NLP research has resulted in the
widespread use of statistical approaches such as machine learning and data mining on a
massive scale. The need for automation is never-ending courtesy of the amount of work
required to be done these days. NLP is a very favorable, but aspect when it comes to
automated applications.
7
Deep learning
Deep learning[108] uses several layers of neurons between the network's inputs and outputs.
The multiple layers can progressively extract higher-level features from the raw input. For
example, in image processing, lower layers may identify edges, while higher layers may
identify the concepts relevant to a human such as digits or letters or faces. [110]
Deep learning has profoundly improved the performance of programs in many important
subfields of artificial intelligence, including computer vision, speech recognition, natural
language processing, image classification[111] and others. The reason that deep learning
performs so well in so many applications is not known as of 2023.[112] The sudden success
of deep learning in 2012–2015 did not occur because of some new discovery or theoretical
breakthrough (deep neural networks and backpropagation had been described by many
people, as far back as the 1950s)[i] but because of two factors: the incredible increase in
computer power (including the hundred-fold increase in speed by switching to GPUs) and
the availability of vast amounts of training data, especially the
giant curated datasets used for benchmark testing, such as
ImageNet.[j] By strict definition, a deep neural network, or DNN,
is a neural network with three or more layers. In practice, most
DNNs have many more layers. DNNs are trained on large
amounts of data to identify and classify phenomena, recognize
patterns and relationships, evaluate possibilities, and make
predictions and decisions. While a single-layer neural network
can make useful, approximate predictions and decisions, the
additional layers in a deep neural network help refine and optimize those outcomes for
greater accuracy. Deep learning drives many
applications and services that improve
automation, performing analytical and
physical tasks without human intervention. It
lies behind everyday products and services—
e.g., digital assistants, voice enabled TV
remotes, credit card fraud detection—as well
as still emerging technologies such as self-
driving cars and generative.AI.
8
Chat GPT
Generative pre-trained transformers (GPT) are large language models that are based on the
semantic relationships between words in sentences (natural language processing). Text
based GPT models are pre-trained on a large corpus of text which can be from the internet.
The pre-training consists in predicting the next token (a token being usually a word, sub
word, or punctuation). Throughout this pre-training, GPT models accumulate knowledge
about the world, and can then generate human-like text by repeatedly predicting the next
token. Typically, a subsequent training phase makes the model more truthful, useful and
harmless, usually with a technique called reinforcement learning from human feedback
(RLHF). Current GPT models are still prone to generating falsehoods called
"hallucinations", although this can be reduced with RLHF and quality data. They are used
in chatbots, which allow you to ask a question or request a task in simple text. Current
models and services include:
Gemini (formerlyBard), ChatGPT, Grok, Claude, Copilot and LLaMA. Multimodal GPT
models can process different types of data (modalities) such as images, videos, sound and
text.
9
Example’s
AI-Powered Medical Imaging Analysis
Background: XYZ Hospital, a leading healthcare institution, is facing challenges in
efficiently analyzing medical images such as X-rays, MRIs, and CT scans. The manual
interpretation of these images by radiologists is time-consuming and can lead to delays in
diagnosis and treatment. To address this issue, the hospital decides to implement an
AIpowered medical imaging analysis system.
Objective: The objective of this case study is to showcase how AI technology can
enhance the efficiency and accuracy of medical image analysis, leading to improved
patient outcomes and operational efficiency for healthcare providers. Implementation:
Algorithm Development:
Data scientists and medical experts collaborate to develop machine learning algorithms
capable of detecting and classifying abnormalities in medical images.
Deep learning techniques, such as convolutional neural networks (CNNs), are utilized due
to their effectiveness in image recognition tasks.
10
Improved Efficiency:
The AI system significantly reduces the time required for image analysis, allowing
radiologists to prioritize complex cases and make quicker treatment decisions.
Turnaround times for reporting are reduced, leading to faster diagnosis and treatment
initiation for patients.
Enhanced Accuracy:
AI algorithms demonstrate high accuracy in detecting abnormalities, including subtle signs
that may be overlooked by human observers.
The combination of AI-driven insights and radiologist expertise leads to more
comprehensive and accurate diagnoses.
Cost Savings:
By streamlining the imaging analysis process, the hospital realizes cost savings through
increased efficiency and reduced reliance on manual labor. Challenges and
Considerations:
Regulatory Compliance:
Compliance with healthcare regulations, such as HIPAA in the United States, requires
careful handling of patient data to ensure privacy and security.
Ethical and Legal Implications:
The use of AI in medical decision-making raises ethical considerations regarding
accountability, transparency, and patient consent.
Clear guidelines and protocols must be established to address these concerns and ensure
responsible AI deployment.
11
References
www.greeksforgreeks.org
https://www.tutorialspoint.com/
https://en.wikipedia.org/
www.slideshare.net
www.copilot.com
12
Conclusion
In conclusion, the case study exemplifies the profound impact of artificial intelligence (AI)
across industries, showcasing its transformative potential in healthcare. Through the
implementation of an AI-powered medical imaging analysis system, XYZ Hospital has
achieved remarkable improvements in efficiency, accuracy, and cost-effectiveness.
13