Home

Awesome

Wireframe

A curated list of papers & resources linked to Wireframe.

Introduction

Contents

2D Wireframe

The wireframe parsing task is firstly proposed by Huang et al. [1]. It proposes the task formulation, a large-scale dataset with wireframe annotations, a baseline method, and a set of evaluation metrics. After that, Zhou et al. [4] and propose an end-to-end solution and significantly improves the performance. Zhang et al. [3] is a counterpart work and refine the annotations of [1]. Meanwhile, Xue et al. [8] is a follow-up work of Xue et al. [2] which holds the new state-of-the-art result. Lin et al. [6] design a hough transform based convolutional operator to treat the line detection task. In order to handle the topology of junction and line, a graph neural network based method LGNN [5] was proposed to resolve the wireframe task. Besides, Huang et al. [7] propose to speed up the line detection. Strictly speaking, line detection is not wireframe parsing as it does not detect junctions of multiple segments.

<p align="center"> <img src="img/demo.gif" width="400"> </p>

The source video above is from kujiale website.

<a name="papers-2d"></a>

Papers

2018

2019

2020

2021

<a name="data-2d"></a>

Datasets (2D)

So far as we know, there exist two wireframe datasets namely ShanghaiTech and YorkUrban.

  1. The ShanghaiTech dataset proposed by Huang et al. [1]. It contains 5,000 training images and 462 test images of man-made scenes which is a basic dataset used by all methods [1-8]. It has another relabeled version SIST which's junctions and lines are carefully labeled.
  2. The YorkUrban is an early existent dataset which is a small dataset containing 102 images and only has Manhattan lines labeled. Most works test their model on it to evaluate the generalizability of models.
<img src="img/ShanghaiTech.png" width="280"><img src="img/YorkUrban.png" width="280">
ShanghaiTech datasetYorkUrban dataset

Leaderboard

DWP[1]AFM[2]PPGNET[3]LCNN[4]LGNN[5]DHT[6]TP-LSD[7]HAWP[8]LETR[9]F-Clip[10]
ShanghaiTechsAP<sup>10</sup>5.124.4-62.962.366.660.666.565.268.3
AP<sup>H</sup>67.869.280.3--84.384.584.785.7
YorkUrbansAP<sup>10</sup>2.19.4-26.4-27.427.428.529.430.8
AP<sup>H</sup>51.048.258.5---60.661.265.0

3D Wireframe

<p align="center"> <img src="img/3d-wireframe.png" width="400"> <img src="img/3d-wireframe.gif" width="400"> </p>

<a name="papers-3d"></a>

Papers (3D)

<a name="data-3d"></a>

Datasets (3D)

So far as we know, there exist two wireframe datasets: one synthetic dataset and one real world dataset. Both of them were proposed by Zhou et al.

  1. SceneCity is synthetic city dataset which is generated through 3D mesh. The dataset includes 230 cities, each containing 8 × 8 city blocks. 100 viewpoints are generated for each city and then render to images through global illumination by Blender modeler. The number of total images is 23000, first 22700 images for training and the rest 300 for validation.

  2. MegaWireframe is a realistic Landmark dataset which is relabeled from MegaDepth v1 dataset. It selects about 200 images from MegaDepth v1 which is manually labeled with wireframes, and registered with the rough 3D depth.

<img src="img/SceneCity.png" width="280"><img src="img/megaDepth.png" width="280">
SceneCityMegaWireframe