Zhong Zhong

Logo


I'm graduate student who has a strong interest in Machine Learning and Data Analysis. I also study Deep Learning (especially, computer vision) during my free time.

笨鸟先飞 & 耐心

My LinkedIn Profile

The is my review of MobileNetV2, 2018 CVPR

A new mobile architecture, MobileNetV2 is introduced here. MobileNetV2 is based on inverted residual structure where the shortcut connections are between the thin bottleneck layers. The intermediate expansion layer inside the inverted residual block uses lightweight depthwise convolutions to filter features as a source of non-linearity. Additionally, the non-linearity of the output layer in the block is removed in order to retain representational power.

This is a followup of the other architecture, MobileNetV1. Modern SOTAs require high computational resources beyond the capbilities of many mobile and embedded applicaitons, thus, MobileNetV2 is introduced to futher overcome this problem beyond MobileNetV1.

1 Inverted residual with linear bottleneck

figure4

1x1 linear bottleneck layer with stride 1

It was found that nonlinearity (ReLu) layer hurts destroy the information perserved from the input. Therefore, a 1x1 linear Conv layer is added at the end of each inverted residual block.

BottleNeck block with stride 2

The block is almost the same to the 1x1 linear bottleneck layer, except that it does not have residual connection and the stride is 2 on the depthwise 3x3 conv layer.

Basic bottleneck convolution

The basic structure of inverted residul is as follows:

table1

2 Overall architecture

table2

3 Ablation Study

4 Experiments