GitHub - uber-research/UPSNet: UPSNet: A Unified Panoptic Segmentation Network
source link: https://github.com/uber-research/UPSNet
Go to the source link to view the article. You can view the picture content, updated content and better typesetting reading experience. If the link is broken, please click the button below to view the snapshot at that time.
README.md
UPSNet: A Unified Panoptic Segmentation Network
Introduction
UPSNet is initially described in a CVPR 2019 oral paper.
Disclaimer
This repository is tested under Python 3.6, PyTorch 0.4.1. And model training is done with 16 GPUs by using horovod. It should also work under Python 2.7 / PyTorch 1.0 and with 4 GPUs.
License
© Uber, 2018-2019. Licensed under the Uber Non-Commercial License.
Citing UPSNet
If you find UPSNet is useful in your research, please consider citing:
@inproceedings{xiong19upsnet,
Author = {Yuwen Xiong, Renjie Liao, Hengshuang Zhao, Rui Hu, Min Bai, Ersin Yumer, Raquel Urtasun},
Title = {UPSNet: A Unified Panoptic Segmentation Network},
Conference = {CVPR},
Year = {2019}
}
Main Results
COCO 2017 (trained on train-2017 set)
test split PQ SQ RQ PQTh PQSt UPSNet-50 val 42.5 78.0 52.4 48.5 33.4 UPSNet-101-DCN test-dev 46.6 80.5 56.9 53.2 36.7
Cityscapes
PQ SQ RQ PQTh PQSt UPSNet-50 59.3 79.7 73.0 54.6 62.7 UPSNet-101-COCO (ms test) 61.8 81.3 74.8 57.6 64.8
Requirements: Software
We recommend using Anaconda3 as it already includes many common packages.
Requirements: Hardware
We recommend using 4~16 GPUs with at least 11 GB memory to train our model.
Installation
Clone this repo to $UPSNet_ROOT
Run init.sh
to build essential C++/CUDA modules and download pretrained model.
For Cityscapes:
Assuming you already downloaded Cityscapes dataset at $CITYSCAPES_ROOT
and TrainIds label images are generated, please create a soft link by ln -s $CITYSCAPES_ROOT data/cityscapes
under UPSNet_ROOT
, and run init_cityscapes.sh
to prepare Cityscapes dataset for UPSNet.
For COCO:
Assuming you already downloaded COCO dataset at $COCO_ROOT
and have annotations
and images
folders under it, please create a soft link by ln -s $COCO_ROOT data/coco
under UPSNet_ROOT
, and run init_coco.sh
to prepare COCO dataset for UPSNet.
Training:
python upsnet/upsnet_end2end_train.py --cfg upsnet/experiments/$EXP.yaml
Test:
python upsnet/upsnet_end2end_test.py --cfg upsnet/experiments/$EXP.yaml
We provide serveral config files (16/4 GPUs for Cityscapes/COCO dataset) under upsnet/experiments folder.
Model & Demo
The model weights that can reproduce numbers in our paper and the demo will be coming soon.
Recommend
About Joyk
Aggregate valuable and interesting links.
Joyk means Joy of geeK