Examples
LeNet
- 1990’s
- 1989
- 1998
AlexNet
2012
- ReLu
- Normalisation
VGG
2015
- 16 layers over AlexNet’s 8
- Looking at vanishing gradient problem
- Xavier
- Similar kernel size throughout
- Gradual filter increase
GoogLeNet
2015
- Inception Layers
- Multiple Loss Functions
Inception Layer
Auxiliary Loss Functions
- Two other SoftMax blocks
- Help train really deep network
- Vanishing gradient problem