Skip to content
/ UDOS Public

UDOS method for Open World Instance Segmentation

License

Notifications You must be signed in to change notification settings

tarun005/UDOS

Folders and files

NameName
Last commit message
Last commit date

Latest commit

625fe23 · Aug 19, 2023

History

15 Commits
Jun 28, 2023
Jul 6, 2023
Jul 6, 2023
Jun 28, 2023
Apr 8, 2023
Jul 24, 2023
Mar 6, 2023
Jun 28, 2023
Jun 28, 2023
Jun 28, 2023
Jul 6, 2023
Mar 6, 2023
Mar 6, 2023
Aug 19, 2023
Jul 6, 2023
Mar 6, 2023
Mar 6, 2023
Mar 6, 2023

Repository files navigation

UDOS: Open-world Instance Segmentation with Bottom-up Supervision

[project page] [paper]

UDOS (Up-Down Open-world Segmentation) is a simple and efficient method for open-world instance segmentation to detect and segment novel objects unseen during training. We leverage bottom-up supervision from unsupervised proposal generation method such as selective search and guide it to learn top-down networks such as MaskRCNN to segment open-world objects not annotated during training. UDOS first predicts part-mask associated with both seen and unseen objects, which is followed by lightweight grouping and refinement modules to predict instance level masks.

Overview of UDOS

Results

Cross-category setting (VOC -> NonVOC)

Model AR100-Box AR100-Mask
LDET 30.9 26.7
OLN 32.5 26.9
GGN 31.6 28.7
UDOS [Ours] 33.5 31.6

More results are provided in our [project page].

Testing Pre-trained models

Cross Category Setting: VOC to NonVOC

Download the pre-trained model from [here].

Then run the inference as follows.

sh jobs/UDOS_test_VOC.sh

Cross Dataset Setting: COCO to UVO

Firstly, download the UVO dataset from [here]. Then, download the pre-trained model trained on complete COCO dataset from [here].

After downloading the pre-trained models, you can run inference using the following script.

sh jobs/UDOS_test_COCO.sh

Training UDOS

Generating super-pixel segmentation.

You can download the super-pixel segmentation from [here]. Alternatively, you can generate your own superpixel segmentation usng selective search as follows.

python -m torch.distributed.launch --nproc_per_node=8 --use_env genSS.py --split train --output_file instances_train2017_SS.json

With the generated/downloaded superpixels, you can use the scripts inside the jobs/ folder to train the UDOS models. For example, to train the cross-category model, you can use

sh jobs/UDOS_train_VOC.sh

Cross-dataset can also be trained similarly, using UDOS_train_COCO.sh. The training should take less than a day on 8 GPUs.

Citation

You can cite our work as follows.

@article{kalluri2023udos
        author    = {Kalluri, Tarun and Wang, Weiyao and Wang, Heng and Chandraker, Manmohan and Torresani, Lorenzo and Tran, Du},
        title     = {Open-world Instance Segmentation: Top-down Learning with Bottom-up Supervision},
        journal   = {arxiv},
        year      = {2023},
      },

About

UDOS method for Open World Instance Segmentation

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published