Need help with MobileNet-V2?
Click the “chat” button below for chat support from the developer who created it, or find similar developers for support.

About the developer

210 Stars 44 Forks Apache License 2.0 75 Commits 3 Opened issues


A Complete and Simple Implementation of MobileNet-V2 in PyTorch

Services available


Need anything else?

Contributors list

# 63,168
64 commits


An implementation of

Google MobileNet-V2
introduced in PyTorch. According to the authors,
improves the state of the art performance of mobile models on multiple tasks and benchmarks. Its architecture is based on an inverted residual structure where the input and output of the residual block are thin bottleneck layers opposite to traditional residual models which use expanded representations in the input an MobileNetV2 uses lightweight depthwise convolutions to filter features in the intermediate expansion layer.

Link to the original paper: Inverted Residuals and Linear Bottlenecks: Mobile Networks for Classification, Detection and Segmentation

This implementation was made to be an example of a common deep learning software architecture. It's simple and designed to be very modular. All of the components needed for training and visualization are added.

Inverted Residuals with Linear Bottlenecks


This project uses Python 3.5.3 and PyTorch 0.3.

Main Dependencies

 pytorch 0.3
 numpy 1.13.1
 tqdm 4.15.0
 easydict 1.7
 matplotlib 2.0.2
 tensorboardX 1.0

Install dependencies:

pip install -r requirements.txt

Train and Test

  1. Prepare your data, then create a dataloader class such as
  2. Create a .json config file for your experiments. Use the given .json config files as a reference.


python config/.json


Due to the lack of computational power. I trained on CIFAR-10 dataset as an example to prove correctness, and was able to achieve test top1-accuracy of 90.9%.

Tensorboard Visualization

Tensorboard is integrated with the project using

library which proved to be very useful as there is no official visualization library in pytorch.

You can start it using:

tensorboard --logdir experimenets//summaries

These are the learning curves for the CIFAR-10 experiment.


Measuring FLOPS on this architecture to compare with other realtime architectures. PyTorch doesn't have a profiler like TensorFlow's. So, I'll be working on measuring FLOPS on my own.


This project is licensed under the Apache License 2.0 - see the LICENSE file for details.

We use cookies. If you continue to browse the site, you agree to the use of cookies. For more information on our use of cookies please see our Privacy Policy.