gll89的个人博客分享 http://blog.sciencenet.cn/u/gll89

博文

[转载]Residual blocks--Building blocks of ResNet

已有 2370 次阅读 2019-4-28 21:15 |个人分类:DeepLearning|系统分类:科研笔记|文章来源:转载

Understanding a residual block is quite easy. In traditional neural networks, each layer feeds into the next layer. In a network with residual blocks, each layer feeds into the next layer and directly into the layers about 2–3 hops away. That’s it. But understanding the intuition behind why it was required in the first place, why it is so important and how similar it looks to some other state of the art architectures is where we are going to focus on. There are more than one interpretations of why residual blocks are awesome and how & why they are one of the key ideas that can make a neural network show state of the art performances on wide range of tasks. Before diving into the details, here is a picture of how a residual block actually looks like.

image.png

We know neural networks are universal function approximators and that the accuracy increases with increasing number of layers. But there is a limit to the number of layers added that result in accuracy improvement. So, if neural networks were universal function approximators then it should have been able to learn any simplex or complex function. But it turns out that, thanks to some problems like vanishing gradients and curse of dimensionality, if we have sufficiently deep networks, it may not be able to learn simple functions like an identity function. Now this is clearly undesirable.



https://blog.sciencenet.cn/blog-1969089-1176021.html

上一篇:common usage of Pandas
下一篇:[转载]Anaconda installation and path configuration on windows
收藏 IP: 60.191.2.*| 热度|

0

该博文允许注册用户评论 请点击登录 评论 (0 个评论)

数据加载中...

Archiver|手机版|科学网 ( 京ICP备07017567号-12 )

GMT+8, 2024-4-25 03:09

Powered by ScienceNet.cn

Copyright © 2007- 中国科学报社

返回顶部