ShuffleNet

2017-10-16  本文已影响23人  信步闲庭v

Approach

The new architecture utilizes two proposed operations, pointwise group convolution and channel shuffle, to greatly reduce computation cost while maintaining accuracy.

We notice 1 × 1 convolutions require(also called pointwise convolutions) considerable complexity. To address the issue, a straightforward solution is to apply channel sparse connections, for example group convolutions, also on 1 × 1 layers. However, this property blocks information flow between channel groups and weakens representation.

If we allow group convolution to obtain input data from different groups, the input and output channels will be fully related. Specifically, for the feature map generated from the previous group layer, we can first divide the channels in each group into several subgroups, then feed each group in the next layer with different subgroups.This can be efficiently and elegantly implemented by a channel shuffle operation.

Experiment

References:
ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices, Xiangyu Zhang

上一篇下一篇

猜你喜欢

热点阅读