Back to Papers
2015|Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun

Deep Residual Learning (ResNet)

Introduces residual learning with skip connections to train very deep networks. The key insight is that learning residual mappings $\mathcal{F}(x) := \mathcal{H}(x) - x$ is easier than learning unreferenced mappings directly. The degradation problem motivated the design: adding more layers to a suitably deep model leads to higher training error. This is not caused by overfitting, but by the difficulty of optimizing deeper networks. Residual connections provide an identity shortcut that allows gradients to flow directly through the network, enabling training of networks with 100+ layers.
Read Original Paper

Implementation Track

4 tasks to complete