Home

Awesome

SCF-Net: Learning Spatial Contextual Features for Large-Scale Point Cloud Segmentation

PWC PWC PWC

architecture

For technical details, please refer to:

SCF-Net: Learning Spatial Contextual Features for Large-Scale Point Cloud Segmentation

(0) Abstract

How to learn effective features from large-scale point clouds for semantic segmentation has attracted increasing attention in recent years. Addressing this problem, we propose a learnable module that learns Spatial Contextual Features from large-scale point clouds, called SCF in this paper. The proposed module mainly consists of three blocks, including the local polar representation block, the dual-distance attentive pooling block, and the global contextual feature block. For each 3D point, the local polar representation block is firstly explored to construct a spatial representation that is invariant to the z-axis rotation, then the dual-distance attentive pooling block is designed to utilize the representations of its neighbors for learning more discriminative local features according to both the geometric and feature distances among them, and finally, the global contextual feature block is designed to learn a global context for each 3D point by utilizing its spatial location and the volume ratio of the neighborhood to the global point cloud. The proposed module could be easily embedded into various network architectures for point cloud segmentation, naturally resulting in a new 3D semantic segmentation network with an encoder-decoder architecture, called SCF-Net in this work. Extensive experimental results on two public datasets demonstrate that the proposed SCF-Net performs better than several state-of-the-art methods in most cases.

abstract

(1) Setup

This code has been tested with Python 3.5, Tensorflow 1.11, CUDA 9.0 and cuDNN v7 on Ubuntu 16.04.

(2) S3DIS

S3DIS

(3) Semantic3D

Semantic3D

(4) STPLS3D

(5) Visualization

S3DIS_VIS

Semantic3D_VIS

python visualization.py --dataset 'xxx' --ply_path "path_to_origin_ply" --label_path "path_to_label"

Citation

If you find our work useful in your research, please consider citing:

@InProceedings{Fan_2021_CVPR,
    author    = {Fan, Siqi and Dong, Qiulei and Zhu, Fenghua and Lv, Yisheng and Ye, Peijun and Wang, Fei-Yue},
    title     = {SCF-Net: Learning Spatial Contextual Features for Large-Scale Point Cloud Segmentation},
    booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
    month     = {June},
    year      = {2021},
    pages     = {14504-14513}}

Acknowledgment

Part of our code refers to the work RandLA-Net