Skip to content

Latest commit

 

History

History
75 lines (53 loc) · 4.09 KB

README.md

File metadata and controls

75 lines (53 loc) · 4.09 KB

Dense Reppoints: Representing Visual Objects with Dense Point Sets

The major contributors include Yinghao Xu*, Ze Yang*, Han Xue*, Zheng Zhang ,Han Hu. (* indicates equal contribution)

This repo is an official implementation of "Dense RepPoints: Representing Visual Objects with Dense Point Sets" on COCO object detection/instance segmentation. The code is based on mmdetection.

Introduction

Dense RepPoints utilizes a dense point set to describe the multi-grained object representation of both box level and pixel level. The following figure illustrates the representation of object segments in different forms using Dense RepPoints. The key techniques to learn such representation are a distance transform sampling (DTS) method combined with a set-to-set supervision method. In inference, both the concave hull and triangulation methods are supported. The method is also efficient, achieving near constant complexity with increasing point number. Please touch arXiv for more details.

Learning Dense RepPoints in Object Detection and Instance Segmentation.

Updates

  • Our paper will appear at ECCV2020(06/07/2020)

Usage

a. Clone the repo, install and download the COCO detection dataset.

git clone --recursive https://github.com/justimyhxu/Dense-RepPoints.git

Please refer to INSTALL.md for installation and dataset preparation in detail.

b. Train with a specific configuration file:

./tools/dist_train.sh ${path-to-cfg-file} ${num_gpu} --validate

Here is one example:

./tools/dist_train.sh configs/dense_reppoints/dense_reppoints_729pts_r50_fpn_1x.py 8 --validate

c. Test script:

./tools/dist_test.sh  ${path-to-cfg-file}  ${model_path} ${num_gpu} --out ${out_file}

Please see GETTING_STARTED.md for the basic usage in detail.

Citing Dense RepPoints

@article{yang2019dense,
  title={Dense reppoints: Representing visual objects with dense point sets},
  author={Yang, Ze and Xu, Yinghao and Xue, Han and Zhang, Zheng and Urtasun, Raquel and Wang, Liwei and Lin, Stephen and Hu, Han},
  journal={arXiv preprint arXiv:1912.11473},
  year={2019}
}

Results and models

The results on COCO 2017val are shown in the table below. More code and models will be added soon.

Method Backbone Anchor convert func Refine Assigner Lr schd box AP mask AP Download
Dense RepPoints R-50-FPN none MinMax MaxIou 1x 39.4 33.8 model
Dense RepPoints R-50-FPN none MinMax ATSS 1x 39.9 33.9 model
Dense RepPoints R-50-FPN none MinMax ATSS 3x (ms-train) 43.4 37.1 model

Notes:

  • R-xx, X-xx denote the ResNet and ResNeXt architectures, respectively.
  • DCN denotes replacing 3x3 conv with the 3x3 deformable convolution in c3-c5 stages of backbone.
  • none in the Anchor column means 2-d center point (x,y) is used to represent the initial object hypothesis.
  • MinMax in the convert func column is the function to convert a point set to a pseudo box.
  • ms-train denotes multi-scale training.
  • ATSS denotes an assigner in Arxiv.