You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Great work here! I have a quick question--
Based on my understanding of your code and the MSRA paper, shouldn't your type B projection shortcuts in the full and bottleneck preactivation models take bn_pre_relu as their input, and not l (excluding the first block)?
As the identity mapping paper states, "For the bottleneck ResNets, when reducing the feature map
size we use projection shortcuts [1] for increasing dimensions, and when pre-
activation is used, these projection shortcuts are also with pre-activation."
Did I misunderstand something? Thanks for your time!
The text was updated successfully, but these errors were encountered:
Hi,
Great work here! I have a quick question--
Based on my understanding of your code and the MSRA paper, shouldn't your type B projection shortcuts in the full and bottleneck preactivation models take
bn_pre_relu
as their input, and notl
(excluding the first block)?As the identity mapping paper states, "For the bottleneck ResNets, when reducing the feature map
size we use projection shortcuts [1] for increasing dimensions, and when pre-
activation is used, these projection shortcuts are also with pre-activation."
Did I misunderstand something? Thanks for your time!
The text was updated successfully, but these errors were encountered: