Skip to content

Crystalxd/DeepLabv3_MobileNetv2_PyTorch

 
 

Repository files navigation

DeepLabv3_MobileNetv2

This is a PyTorch implementation of MobileNet v2 network with DeepLab v3 structure used for semantic segmentation.

The backbone of MobileNetv2 comes from paper:

Inverted Residuals and Linear Bottlenecks: Mobile Networks for Classification, Detection and Segmentation

And the segment head of DeepLabv3 comes from paper:

Rethinking Atrous Convolution for Semantic Image Segmentation

Please refer to these papers about details like Atrous Convolution, Inverted Residuals, Depthwise Convolution or ASPP if you have some confusion about these blocks.

Results

After training for 150 epochs, without any further tuning, the first training result on test set is like: img

Feel free to change any config or code in this repo :-)

How to use?

First you need to install dependencies of this implementation. This implementation is written under Python 3.5 with following libs:

torch 0.4.0
torchvision 0.2.1
numpy 1.14.5
opencv-python 3.4.1.15
tensorflow 1.8.0 (necessary for tensorboardX)
tensorboardX 1.2

use sudo pip install lib to install them

Then, prepare cityscapes dataset or your own dataset. Currently, cityscapes is the only supported dataset without any modification.

Cityscapes dataset should have the following hierachy:

dataset_root
|   trainImages.txt
|   trainLabels.txt
|   valImages.txt
|   valLabels.txt 
|
└───gtFine(Label Folder)
|   └───train(train set)
|   |   └───aachen(city)
|   |   └───bochum
|   |   └───...
|   |
|   └───test(test set)
|   └───val(val set)
|
└───leftImg8bit(Image Folder)
    └───train
    └───test
    └───val

Don't worry about txt files if you don't have them, this program can generate unexist txt files automatically.

Third, modify config.py to fit your own training policy or configuration

At last, run python main.py --root /your/path/to/dataset/ or just run python main.py

After training, tensorboard is also available to observe training procedure using tensorboard --logdir=./exp_dir/summaries

Tips

I have changed a little from origin MobileNetv2 and DeepLabv3 network, here are the changes:

1. The multi-grid blocks have the same structure with the 7-th layer in MobileNetv2 while 
the rest layers of MobileNetv2 are discarded.
2. The lr decay is determined by epoch not iterations as in DeepLab and the input image 
is randomly cropped by 512 instead of 513 in DeepLab.
3. During training, a input image is first resized so that the shorter side is 600 pixel, 
then cropped into 512 pixels square and sent into network.

If you have some question, please leave an issue.

ImageNet pre-trained weights are loaded from Randl's github, really helpful!

TO-DO

  1. add cityscapes visualization tools(Done)
  2. fine-tune training policy
  3. use ImageNet pre-trained model(Done)

Logs

Date Changes
7.11 fix bugs in network.Test(), add cityscapes output visualization function
7.12 fix bugs in network.plot_curve(), add checkpoint split to avoid out of memory, add save loss in network.save_checkpoint()
7.13 fix bugs in figure save, add checkpoint@150epoch
7.27 upload ImageNet pre-trained weight

About

A PyTorch Implementation of MobileNetv2+DeepLabv3

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%