Lecture 9 | CNN Architectures

2019-11-04  本文已影响0人  Ysgc
left: setting; right: tricks on each gpu, there's only one half of all the neurons

training classification early -> more gradient in such a deep NN

[图片上传失败...(image-c8f98-1572805673386)]

from test error -> deeper is worse -> overfitting ???
from training error -> deeper is worse -> not overfitting !!! -> optimization problem ??? -> deeper means harder to train ???

cifar -> smaller than imagenet

stack MLP on top of the conv

local network -> precursor of googlenet

上一篇 下一篇

猜你喜欢

热点阅读