jmlr.org
scholar.google.com
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
Ioffe, Sergey and Szegedy, Christian
International Conference on Machine Learning - 2015 via Local Bibsonomy
Keywords: dblp


[link]
Summary by José Manuel Rodríguez Sotelo 8 years ago
Loading...
Your comment:
[link]
Summary by Shagun Sodhani 8 years ago
Loading...
Do you have a source for how the normalization works for CNNs? Do you know of any follow-up work which did what you mentioned in "Future work"? (And there is a typo: "archwitecture")

To see effect of batch normalization on CNN, you may refer this benchmark [https://github.com/ducha-aiki/caffenet-benchmark/blob/master/batchnorm.md] Thanks for pointing out the typo :)

Your comment:
[link]
Summary by Alexander Jung 7 years ago
Loading...
Your comment:
[link]
Summary by Denny Britz 8 years ago
Loading...
Could you please explain why adding the parameters $\beta$ and $\gamma$ does not change the variance?

What do you mean by "shuffle training examples more thoroughly"?

Your comment:
[link]
Summary by Martin Thoma 8 years ago
Loading...
Your comment:
[link]
Summary by Cubs Reading Group 7 years ago
Loading...
Your comment:
[link]
Summary by Léo Paillier 6 years ago
Loading...
Your comment:
[link]
Summary by Joseph Paul Cohen 8 years ago
Loading...
Your comment:


ShortScience.org allows researchers to publish paper summaries that are voted on and ranked!
About

Sponsored by: