Professional Documents
Culture Documents
Types of Data:: Reference Website
Types of Data:: Reference Website
org/difference-between-machine-learning-and-
artificial-intelligence/
https://www.interviewbit.com/blog/artificial-intelligence-and-machine-
learning/
https://en.wikipedia.org/wiki/Deep_learning#/media/File:AI-ML-DL.svg
https://www.simplilearn.com/tutorials/artificial-intelligence-tutorial/ai-vs-
machine-learning-vs-deep-learning
Types of Data :
https://www.mygreatlearning.com/blog/types-of-data/
Types of Training :
https://www.geeksforgeeks.org/ml-types-learning-supervised-learning/?
ref=lbp
Reference website :
https://www.techtarget.com/whatis/definition/validation-set#:~:text=A
%20validation%20set%20is%20a,a%20corpus%20of%20training
%20data.
A validation set is a set of data used to train artificial intelligence (AI) with
the goal of finding and optimizing the best model to solve a given
problem. Validation sets are also known as dev sets.
A supervised AI is trained on a corpus of training data. Training, tuning,
model selection and testing are performed with three different datasets:
the training set, the validation set and the testing set. Validation sets are
used to select and tune the final AI model.
The validation set makes up about 20 percent of the bulk of data used.
The validation set contrasts with training and test sets in that it is an
intermediate phase used for choosing the best model and optimizing it.
Validation is sometimes considered a part of the training phase. It is in
this phase that parameter tuning occurs for optimizing the selected
model. Testing sets make up 20 percent of the bulk of the data. These
sets are ideal data and results with which to verify correct operation of
an AI. The test set is ensured to be the input data grouped together with
verified correct outputs, generally by human verification. This ideal set is
used to test results and assess the performance of the final model.
Reinforcement learning:
https://www.geeksforgeeks.org/what-is-reinforcement-learning/
https://www.geeksforgeeks.org/understanding-reinforcement-learning-in-
depth/
Types of Reinforcement:
There are two types of Reinforcement:
1. Positive: Positive Reinforcement is defined as when an event,
occurs due to a particular behavior, increases the strength and the
frequency of the behavior. In other words, it has a positive effect on
behavior.
Advantages of reinforcement learning are:
Maximizes Performance
Sustain Change for a long period of time
Too much Reinforcement can lead to an overload of states which
can diminish the results
2. Negative: Negative Reinforcement is defined as strengthening of
behavior because a negative condition is stopped or avoided.
Advantages of reinforcement learning:
Increases Behavior
Provide defiance to a minimum standard of performance
It Only provides enough to meet up the minimum behavior
Deep Learning :
Reference: https://www.javatpoint.com/deep-learning
https://www.javatpoint.com/machine-learning-vs-deep-learning
A Perceptron is an Artificial Neuron. It is the simplest possible Neural Network.
Neural Networks are the building blocks of Machine Learning.
Deep learning is a branch of machine learning which is based on artificial
neural networks. It is capable of learning complex patterns and relationships
within data. In deep learning, we don’t need to explicitly program everything.
It has become increasingly popular in recent years due to the advances in
processing power and the availability of large datasets. Because it is based
on artificial neural networks (ANNs) also known as deep neural networks
(DNNs). These neural networks are inspired by the structure and function of
the human brain’s biological neurons, and they are designed to learn from
large amounts of data.
The algorithms used in facial recognition are one instance of bias in machine
learning. According to research, facial recognition software performs worse on
those with darker skin tones, which causes false positive and false negative
rates to be higher for people of races. This bias may have significant
consequences, particularly in law enforcement and security applications,
where false positives may result in unjustified arrests or other undesirable
results.
Researchers are looking into novel techniques for creating synthetic data that
may be used to supplement small datasets to address this constraint. To
expand the amount of data accessible for training machine learning algorithms,
efforts are also being made to enhance data sharing and collaboration across
enterprises.
A major obstacle to machine learning is the need for more available data.
Addressing this restriction will need for a concerted effort across industries and
disciplines to improve data collection, sharing, and reinforcement in order to
ensure that machine learning algorithms can continue to be helpful in a variety
of applications.
5. Computational Resources
Machine learning algorithms can be computationally expensive, and they may
require a lot of resources to be successfully trained. This may be a major
barrier, particularly for people or smaller companies who want access to high-
performance computing resources. Distributed and cloud computing can be
used to get around this restriction, however the project's cost might go up.
For huge datasets and complex models, machine learning approaches can be
computationally expensive. The scalability and feasibility of machine learning
algorithms may be hampered by the need for significant processing resources.
The availability of computational resources like processor speed, memory, and
storage is another limitation on machine learning.
6. Lack of Causality
Predictions based on correlations in the data are frequently made using
machine learning algorithms. Machine learning algorithms may not shed light
on the underlying causal links in the data because correlation does not always
imply causation. This may reduce our capacity for precise prediction when
causality is crucial.
The absence of causation is one of machine learning's main drawbacks. The
main purpose of machine learning algorithms is to find patterns and
correlations in data; however, they cannot establish causal links between
different variables. In other words, machine learning models can forecast
future events based on seen data, but they cannot explain why such events
occur.
To get over this restriction, machine learning may need to be integrated with
other methodologies like experimental design. Researchers can identify causal
relationships by manipulating variables and observing how those changes
impact a result using an experimental design. However, compared to
traditional machine learning techniques, this approach may require more time
and resources.
Machine learning can be a useful tool for predicting outcomes from observable
data, but it's crucial to be aware of its limitations when making decisions based
on these predictions. The lack of causation is a basic flaw in machine learning
systems. To establish causation, it could be necessary to use methods other
than machine learning.
7. Ethical Considerations
Machine learning models can have major social, ethical, and legal
repercussions when used to make judgments that affect people's lives.
Machine learning models, for instance, may have a differential effect on groups
of individuals when used to make employment or lending choices. Privacy,
security, and data ownership must also be addressed when adopting machine
learning models.
The ethical issue of bias and discrimination is a major one. If the training data is
biased or the algorithms are not created in a fair and inclusive manner, biases
and discrimination in society may be perpetuated and even amplified by
machine learning algorithms.
Finally, there are ethical issues around how machine learning will affect
society. More sophisticated machine learning algorithms may have far-
reaching social, economic, and political repercussions that require careful
analysis and regulation