Residual Network (ResNet)
ResNet makes it possible to train up to hundreds or even thousands of layers and still achieves compelling performance. Thanks to this technique they were able to train a network with 152 layers while still having lower complexity than VGGNet. It achieves a top-5 error rate of 3.57%
Evolution of CNN Architectures: LeNet, AlexNet, ZFNet, GoogleNet, VGG and ResNet
It all started with LeNet in 1998 and eventually, after nearly 15 years, lead to ground breaking models winning the ImageNet Large Scale Visual Recognition Challenge which includes AlexNet in 2012, ZFNet in 2013, GoogleNet in 2014, VGG in 2014, ResNet in 2015 to ensemble of previous models in 2016.