Professional Documents
Culture Documents
Goog Le Net
Goog Le Net
Goog Le Net
Researchers noted the as we increase the number of convolutional layer the results are
getting batter But as you can imagine, this can often create complications like:
The bigger the model, the more prone it is to overfitting. This is particularly noticeable
when the training data is small.
GoogLeNet
Motivation (Going Deeper)
Researchers noted the as we increase the number of convolutional layer the results are
getting batter But as you can imagine, this can often create complications:
The bigger the model, the more prone it is to overfitting. This is particularly noticeable
when the training data is small.
Increasing the number of parameters means you need to increase your existing
computational resources.
GoogLeNet
Motivation (Going Deeper)
Researchers noted the as we increase the number of convolutional layer the results are
getting batter But as you can imagine, this can often create complications:
The bigger the model, the more prone it is to overfitting. This is particularly noticeable
when the training data is small.
Increasing the number of parameters means you need to increase your existing
computational resources.
The problems like vanishing gradient can occur which training very deep models.
Going Deeper and Wider
Addressed the complexity problem of architecture by forming
wider network instead of deeper by introducing the concept of
inception module.
Going Deeper and Wider
(Naïve Inception)
Addressed the complexity problem of architecture by forming
wider network instead of deeper by introducing the concept of
inception module.
First inception module called naïve inception module
incorporated 3 convolutions with different kernel or filter sizes
(1x1, 3x3 and 5x5) at the same level and a 3x3 max pooling.
Outputs of these 3 convolutions and pooling are concatenated
and fed to the next inception module.
Naïve Inception
It is already a very deep model compared with previous AlexNet, ZFNet and
VGGNet.
There are numerous inception modules connected to go deeper.
Global Average Pooling
Previously fully connected layers were used at the end of architecture which adds
a lot of parameters and increases computational complexity But in GoogLeNet
global average pooling is introduced.
Global Average Pooling
1. Reduced number of operations and computational cost because of the use of 1x1
convolution.
GoogLeNet
Pros:
1. Reduced number of operations and computational cost because of the use of 1x1
convolution.
2. Capable of capturing features at multiple scales with the help of different kernel size;
3x3 and 5x5, at the same level.
GoogLeNet
Pros:
1. Reduced number of operations and computational cost because of the use of 1x1
convolution.
2. Capable of capturing features at multiple scales with the help of different kernel size;
3x3 and 5x5, at the same level.
3. Also performs well while working with low contrast images.
4. Highly scalable according to available resources because of its modular architecture.
GoogLeNet
Pros:
1. Reduced number of operations and computational cost because of the use of 1x1
convolution.
2. Capable of capturing features at multiple scales with the help of different kernel size;
3x3 and 5x5, at the same level.
3. Also performs well while working with low contrast images.
4. Highly scalable according to available resources because of its modular architecture.
5. Avoid vanishing gradient problem to some extent by introducing the auxiliary classifiers.
6. Requires less memory compared to previous models like VGG and AlexNet.
GoogLeNet
Cons:
Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., ... & Rabinovich, A. (2015).
Going deeper with convolutions. In Proceedings of the IEEE conference on computer vision
and pattern recognition (pp. 1-9).