Paying more attention to attenti

2017-10-20  本文已影响112人  信步闲庭v

Approach

By properly defining attention for convolutional neural networks, we can actually use this type of information in order to significantly improve the performance of a student CNN network by forcing it to mimic the attention maps of a powerful teacher network.

To define a spatial attention mapping function, the implicit assumption that we make is that the absolute value of a hidden neuron activation (that results when the network is evaluated on given input) can be used as an indication about the importance of that neuron w.r.t. the specific input.


Mid-level attention maps have higher activation level around eyes, nose and lips, high-level activations correspond to the whole face.

Experiment

References:
PAYING M ORE A TTENTION TO A TTENTION :I MPROVING THE P ERFORMANCE OF C ONVOLUTIONALN EURAL N ETWORKS VIA A TTENTION TRANSFER, Sergey Zagoruyko, 2017, ICLR

上一篇 下一篇

猜你喜欢

热点阅读