Professional Documents
Culture Documents
AI Project Cycle ppt_Notes
AI Project Cycle ppt_Notes
A step-by-step process that a person should follow to develop an AI Project to solve a problem. AI
Project Cycle provides us with an appropriate framework which can lead us to achieve our goal.
• Problem Scoping
• Data Acquisition
• Data Exploration
• Modelling
• Evaluation
Problem Scoping
Identifying a problem and having a vision to solve it, is called Problem Scoping. Scoping a problem is
not that easy as we need to have a deeper understanding so that the picture becomes clearer while
we are working to solve it. So we use the 4Ws Problem Canvas to understand the problem in a better
way.
Problem Canvas
The 4Ws Problem canvas helps in identifying the key elements related to the problem. The 4Ws are :
Who
What
Where
Why
1. Who : This block helps in analysing the people who are getting affected directly or indirectly due
to a problem. Under this, we find out who are the ‘Stakeholders’ (those people who face this
problem and would be benefitted with the solution) to this problem? Below are the questions that
we need to discuss under this block.
3. Where: This block will help us to look into the situation in which the problem arises, the context of
it, and the locations where it is prominent. Here is the Where Canvas:
4. Why: In the “Why” canvas, we think about the benefits which the stakeholders would get from the
solution and how it will benefit them as well as the society. Below are the questions that we need to
discuss under this block.
This is the second stage of AI Project cycle. According to the term, this stage is about acquiring data
for the project. Whenever we want an AI project to be able to predict an output, we need to train it
first using data.
For example, If you want to make an Artificially Intelligent system which can predict the salary of any
employee based on his previous salaries, you would feed the data of his previous salaries into the
machine. The previous salary data here is known as Training Data while the next salary prediction
data set is known as the Testing Data.
Data features refer to the type of data you want to collect. In above example, data features would be
salary amount, increment percentage, increment period, bonus, etc. There can be various ways to
collect the data. Some of them are:
Surveys
Web Scraping
Sensors
Cameras
Observations
API (Application Program Interface)
One of the most reliable and authentic sources of information, are the open-sourced websites
hosted by the government. Some of the open-sourced Govt. portals are: data.gov.in, india.gov.in
Quickly get a sense of the trends, relationships and patterns contained within the data.
Define strategy for which model to use at a later stage.
Communicate the same to others effectively.
To visualise data, we can use various types of visual representations like Bargraph,
Histogram, Line Chart, Pie Chart.
The graphical representation makes the data understandable for humans as we can discover trends
and patterns out of it, but machine can analyse the data only when the data is in the most basic form
of numbers (which is binary – 0s and 1s). The ability to mathematically describe the relationship
between parameters is the heart of every AI model.
It refers to the AI modelling where the rules are defined by the developer. The machine follows the
rules or instructions mentioned by the developer and performs its task accordingly.
In this we fed the data along with rules to the machine and the machine after getting trained on
them is now able to predict answers for the same. A drawback/feature for this approach is that the
learning is static.
1. A rule-based system is built upon a lot of data, deep knowledge of the domain, and a lot of
manual work.
2. Writing and generating rules for a complex system is quite challenging and time-consuming.
3. The self-learning capacity in a rule-based system is less as it generates the result as per the
rules.
4. Complex pattern identification is a challenging task in the Rule Based method as it takes a lot
of time and analysis.
It refers to the AI modelling where the machine learns by itself. In this approach the AI model gets
trained on the data fed to it and then is able to design a model which is adaptive to the change in
data. An advantage for this approach is that the learning is dynamic. The learning-based approach
can further be divided into three parts:
Supervised Learning: In a supervised learning model, the dataset which is fed to the machine is
labelled. A label is some information which can be used as a tag for data. For example, students get
grades according to the marks they secure in examinations. These grades are labels which categorise
the students according to their marks. There are two types of Supervised Learning models:
Classification: Where the data is classified according to the labels. This model works on
discrete dataset which means the data need not be continuous.
The best example to understand the Classification problem is Email Spam Detection. The
model is trained on the basis of millions of emails on different parameters, and whenever it
receives a new email, it identifies whether the email is spam or not. If the email is spam,
then it is moved to the Spam folder.
Regression: Such models work on continuous data. For example, if we wish to predict our
next salary, then we would put in the data of our previous salary, any increments, etc., and
would train the model. Here, the data which has been fed to the machine is continuous.
The task of the Regression algorithm is to find the mapping function to map the input
variable(x) to the continuous output variable(y).
Example: Suppose we want to do weather forecasting, so for this, we will use the Regression
algorithm. In weather prediction, the model is trained on the past data, and once the
training is completed, it can easily predict the weather for future days.
In Regression, the output variable must In Classification, the output variable must be a discrete
be of continuous nature or real value. value.
The task of the regression algorithm is to The task of the classification algorithm is to map the
map the input value (x) with the input value(x) with the discrete output variable(y).
continuous output variable(y).
Regression Algorithms are used with Classification Algorithms are used with discrete data.
continuous data.
In Regression, we try to find the best fit In Classification, we try to find the decision boundary,
line, which can predict the output more which can divide the dataset into different classes.
accurately.
The regression Algorithm can be further The Classification algorithms can be divided into Binary
divided into Linear and Non-linear Classifier and Multi-class Classifier.
Regression.
Unsupervised Learning: An unsupervised learning model works on unlabelled dataset. This means
that the data which is fed to the machine is random. This model is used to identify relationships,
patterns and trends out of the data which is fed into it. It helps the user in understanding what the
data is about and what are the major features identified by the machine in it.
Clustering: Clustering or cluster analysis is a machine learning technique, which groups the
unlabelled dataset. It can be defined as "A way of grouping the data points into different
clusters, consisting of similar data points. The objects with the possible similarities remain
in a group that has less or no similarities with another group."
It does it by finding some similar patterns in the unlabelled dataset such as shape, size, color,
behavior, etc., and divides them as per the presence and absence of those similar patterns.
It is an unsupervised learning method, hence no supervision is provided to the algorithm,
and it deals with the unlabeled dataset.
After applying this clustering technique, each cluster or group is provided with a cluster-ID.
ML system can use this id to simplify the processing of large and complex datasets.
The clustering technique can be widely used in various tasks. Some most common uses of
this technique are:
o Market Segmentation
o Statistical data analysis
o Social network analysis
o Image segmentation
o Anomaly detection, et
Some benefits of applying dimensionality reduction technique to the given dataset are given
below:
o By reducing the dimensions of the features, the space required to store the dataset also gets
reduced.
o Reduced dimensions of features of the dataset help in visualizing the data quickly.
There are also some disadvantages of applying the dimensionality reduction, which are given
below:
1. Robotics: Robots with pre-programmed behavior are useful in structured environments, such as
the assembly line of an automobile manufacturing plant, where the task is repetitive in nature.
2. A master chess player makes a move. The choice is informed both by planning, anticipating
possible replies and counter replies.
3. The only way to collect information about the environment is to interact with it.
1. Reinforcement learning can be used to solve very complex problems that cannot be solved by
conventional techniques.
2. The model can correct the errors that occurred during the training process.
3. In RL, training data is obtained via the direct interaction of the agent with the environment
4. Reinforcement learning can handle environments that are non-deterministic, meaning that the
outcomes of actions are not always predictable. This is useful in real-world applications where the
environment may change over time or is uncertain.
5. Reinforcement learning can be used to solve a wide range of problems, including those that
involve decision making, control, and optimization.
6. Reinforcement learning is a flexible approach that can be combined with other machine learning
techniques, such as deep learning, to improve performance.
3. Reinforcement learning is highly dependent on the quality of the reward function. If the reward
function is poorly designed, the agent may not learn the desired behavior.
4. Reinforcement learning can be difficult to debug and interpret. It is not always clear why the agent
is behaving in a certain way, which can make it difficult to diagnose and fix problems.
What is Evaluation
Once a model has been made and trained, it needs to go through proper testing so that one can
calculate the efficiency and performance of the model. Hence, the model is tested with the help of
Testing Data and the efficiency of the model is calculated on the basis of the parameters mentioned
below
Accuracy
Precision
Recall
F1 Score (F1 score is a machine learning evaluation metric that measures a model's accuracy)
Neural Network
Neural networks are loosely modelled after how neurons in the human brain behave. The key
advantage of neural networks is that they are able to extract data features automatically without
needing the input of the programmer. It is a fast and efficient way to solve problems for which the
dataset is very large, such as in images.
As seen in the figure given above, the larger Neural Networks tend to perform better with larger
amounts of data whereas the traditional machine learning algorithms stop improving after a certain
saturation point.
A Neural Network is divided into multiple layers and each layer is further divided into several blocks
called nodes. The first layer of a Neural Network is known as the input layer. Its job is to acquire data
and feed it to the Neural Network. No processing occurs at the input layer. Next to it, are the hidden
layers. Hidden layers are the layers in which the whole processing occurs. These layers are hidden
and are not visible to the user. There can be multiple hidden layers in a neural network system. The
last hidden layer passes the final processed data to the output layer which then gives it to the user as
the final output.
Neural Network Systems are modelled on the human brain and nervous system.
They are able to automatically extract features without input from the programmer.
Every neural network node is essentially a machine learning algorithm.
It is useful when solving problems for which the data set is very large.