Professional Documents
Culture Documents
ML Lab Notes
ML Lab Notes
ML Lab Notes
recall refers to percentage of total relevant results correctly classified by our algo
true positive
true positive + false negative
Sensitivity
measure proportion of actual positive cases that got predicted as positive
MLP(Multilayer Perceptron)
class of feed forward artificial neural network
it utilises a supervised learning technique called backpropagation for training
Activation function
decides whether a neuron shouls be activated or not ( decides whether the neurons input is
important to the network or not in the process of prediction using simpler mathematical operations.)
Backpropagation
used to change the weights of neural net based on error rate obtained in previous epoch.
Forward Propagation
in neural networks we forward propagate to get the output and compare it with the real value to get
the error
Perceptron
single layer neural network
Weights
shows the strength of particular node
Bias
allows to shift the activation function curve up and down
Steps-
Standardisation
transform all variables to one scale. It reduces the biasness among variables
Z= value-mean /
standard deviation
feature vector
choose whether to keep all components or discrad the ones of lesser significance
Statistical Testing
determine whether the random variable following null hypothesis or alternate hypotheseis
Null hypothesis- there is no significance difference between sample and population or among
different populations.
Hypothesis testing
evaluates the evidence data provides sgainst a hypothesis
T-test
used to compare means of two given samples
F-Test
used to compare standard deviation of two samples.
Prunning
Data compression technique that reduces the size of decision tree by removing sections of tree that
are non-critical and redundant to classify instances.
for each non-leaf node in the tree, temporarily prune the tree below and then test the accuracy of
hypothesis on validation set.
If the accuracy increases permanently prune the node.
Post pruning
grow full tree and then remove nodes
Pre pruning
stop growing when data split not statistically significant.
Entropy
measures uncertainity, purity and informatioon content
why decision tree is supervised?
SVM
supervised model used for classification and regression problems
works well when there is understandable margin of dissociation between classes
more productive in high dimensional spaces
Kernel trick
is a simple method where a non linear data is projected onto a higher dimension space so as to make
it easier to classify the data where it could be linearly divided by a plane.
Kernel Function
is a method to take data as input and transform into the required form of processing data.
Reinforcement Learning
training method based on rewarding desired behaviours and/or punishing undesired ones.
Learning agent interpret the environment, take actions and learn through trial and error.
Poisson Distribution
measures the probability of a given number of events happening in a specified time period
Random Forest
supervised learning
used for both regression and classification
it builds multiple decision trees and merges them together to get a more accurate and stable
prediction
Bagging
used to reduce variance within noisy dataset
Bagging: It is a homogeneous weak learners’ model that learns from each other independently
in parallel and combines them for determining the model average.
1.Boosting: It is also a homogeneous weak learners’ model but works differently from
Bagging. In this model, learners learn sequentially and adaptively to improve model
predictions of a learning algorithm.